CN116363123A - Fluorescence microscopic imaging system and method for detecting circulating tumor cells - Google Patents
Fluorescence microscopic imaging system and method for detecting circulating tumor cells Download PDFInfo
- Publication number
- CN116363123A CN116363123A CN202310581412.8A CN202310581412A CN116363123A CN 116363123 A CN116363123 A CN 116363123A CN 202310581412 A CN202310581412 A CN 202310581412A CN 116363123 A CN116363123 A CN 116363123A
- Authority
- CN
- China
- Prior art keywords
- fluorescence
- training
- feature map
- shallow
- map
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 208000005443 Circulating Neoplastic Cells Diseases 0.000 title claims abstract description 42
- 238000000799 fluorescence microscopy Methods 0.000 title claims abstract description 39
- 238000000034 method Methods 0.000 title claims abstract description 33
- 230000011218 segmentation Effects 0.000 claims abstract description 63
- 238000011161 development Methods 0.000 claims abstract description 60
- 230000018109 developmental process Effects 0.000 claims abstract description 60
- 238000007781 pre-processing Methods 0.000 claims abstract description 33
- 238000003062 neural network model Methods 0.000 claims abstract description 17
- 238000012549 training Methods 0.000 claims description 170
- 238000013527 convolutional neural network Methods 0.000 claims description 75
- 239000011159 matrix material Substances 0.000 claims description 61
- 238000001514 detection method Methods 0.000 claims description 32
- 238000011176 pooling Methods 0.000 claims description 32
- 238000003384 imaging method Methods 0.000 claims description 30
- 238000010586 diagram Methods 0.000 claims description 21
- 238000000605 extraction Methods 0.000 claims description 21
- 238000012545 processing Methods 0.000 claims description 16
- 230000004927 fusion Effects 0.000 claims description 15
- 238000005457 optimization Methods 0.000 claims description 15
- 230000004913 activation Effects 0.000 claims description 12
- 238000005728 strengthening Methods 0.000 claims description 10
- 238000012546 transfer Methods 0.000 claims description 6
- 230000006870 function Effects 0.000 description 20
- 238000004590 computer program Methods 0.000 description 6
- 230000000694 effects Effects 0.000 description 6
- 210000004027 cell Anatomy 0.000 description 5
- 238000010606 normalization Methods 0.000 description 5
- 238000003860 storage Methods 0.000 description 5
- 239000013598 vector Substances 0.000 description 5
- 210000000601 blood cell Anatomy 0.000 description 4
- 230000007246 mechanism Effects 0.000 description 4
- 206010028980 Neoplasm Diseases 0.000 description 3
- 238000004458 analytical method Methods 0.000 description 3
- 230000008878 coupling Effects 0.000 description 3
- 238000010168 coupling process Methods 0.000 description 3
- 238000005859 coupling reaction Methods 0.000 description 3
- 230000008569 process Effects 0.000 description 3
- 210000004369 blood Anatomy 0.000 description 2
- 239000008280 blood Substances 0.000 description 2
- 238000004891 communication Methods 0.000 description 2
- 238000012937 correction Methods 0.000 description 2
- 238000013135 deep learning Methods 0.000 description 2
- 238000013399 early diagnosis Methods 0.000 description 2
- 238000012544 monitoring process Methods 0.000 description 2
- 238000000926 separation method Methods 0.000 description 2
- 239000000427 antigen Substances 0.000 description 1
- 102000036639 antigens Human genes 0.000 description 1
- 108091007433 antigens Proteins 0.000 description 1
- 238000013528 artificial neural network Methods 0.000 description 1
- 230000005540 biological transmission Effects 0.000 description 1
- 230000017531 blood circulation Effects 0.000 description 1
- 238000012512 characterization method Methods 0.000 description 1
- 238000013500 data storage Methods 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 238000011156 evaluation Methods 0.000 description 1
- 230000005284 excitation Effects 0.000 description 1
- 239000000835 fiber Substances 0.000 description 1
- 230000010354 integration Effects 0.000 description 1
- 230000001678 irradiating effect Effects 0.000 description 1
- 238000002955 isolation Methods 0.000 description 1
- 238000012886 linear function Methods 0.000 description 1
- 206010061289 metastatic neoplasm Diseases 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 230000000737 periodic effect Effects 0.000 description 1
- 230000000704 physical effect Effects 0.000 description 1
- 238000004393 prognosis Methods 0.000 description 1
- 230000002787 reinforcement Effects 0.000 description 1
- 230000003252 repetitive effect Effects 0.000 description 1
- 238000005096 rolling process Methods 0.000 description 1
- 238000005070 sampling Methods 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 239000007787 solid Substances 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/0002—Inspection of images, e.g. flaw detection
- G06T7/0012—Biomedical image inspection
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/26—Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10056—Microscopic image
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10064—Fluorescence image
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20081—Training; Learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20084—Artificial neural networks [ANN]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30004—Biomedical image processing
- G06T2207/30096—Tumor; Lesion
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02A—TECHNOLOGIES FOR ADAPTATION TO CLIMATE CHANGE
- Y02A90/00—Technologies having an indirect contribution to adaptation to climate change
- Y02A90/10—Information and communication technologies [ICT] supporting adaptation to climate change, e.g. for weather forecasting or climate simulation
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02T—CLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
- Y02T10/00—Road transport of goods or passengers
- Y02T10/10—Internal combustion engine [ICE] based vehicles
- Y02T10/40—Engine management systems
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- General Health & Medical Sciences (AREA)
- Health & Medical Sciences (AREA)
- Artificial Intelligence (AREA)
- Computational Linguistics (AREA)
- Quality & Reliability (AREA)
- Radiology & Medical Imaging (AREA)
- Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
- Life Sciences & Earth Sciences (AREA)
- Medical Informatics (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Data Mining & Analysis (AREA)
- Evolutionary Computation (AREA)
- Molecular Biology (AREA)
- Computing Systems (AREA)
- General Engineering & Computer Science (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Multimedia (AREA)
- Image Analysis (AREA)
Abstract
The utility model relates to the technical field of fluorescence microscopic imaging, and specifically discloses a fluorescence microscopic imaging system for detecting circulating tumor cells and a method thereof, wherein a fluorescence development image of a detected sample is firstly obtained, then the fluorescence development image is subjected to image preprocessing and passes through a shallow feature extractor based on a first convolution neural network model to obtain a fluorescence shallow feature map, then the fluorescence shallow feature map passes through a spatial attention module and a deep feature extractor based on a second convolution neural network model to obtain a fluorescence deep feature map, finally, the spatial enhanced fluorescence shallow feature map and the fluorescence deep feature map are fused, then image semantic segmentation is carried out through a decoder to obtain a segmentation result, and the number of CTC cells is determined based on the segmentation result. In this way, the number of CTC cells is accurately calculated.
Description
Technical Field
The present application relates to the field of fluorescence microscopy imaging technology, and more particularly, to a fluorescence microscopy imaging system for detecting circulating tumor cells and a method thereof.
Background
Circulating Tumor Cells (CTCs) are rare cells shed into the blood circulation from primary or metastatic tumors and can serve as important markers for tumor biology and clinical treatment. The detection and analysis of CTCs are of great significance for early diagnosis, prognosis evaluation, treatment effect monitoring and personalized treatment of tumors.
However, the detection and isolation of CTCs is a very challenging task because of the extremely low number of CTCs in the blood (about 1-10 per milliliter of blood) and the morphology and size that is similar to normal blood cells. At present, the commonly used CTC detection methods mainly comprise a surface antigen-based immune enrichment method, a physical property-based separation method, a functional property-based separation method and the like, but the methods have certain limitations such as low specificity, complex operation, large sample loss and the like.
Thus, an optimized solution is desired.
Disclosure of Invention
The application provides a fluorescence microscopic imaging system for detecting circulating tumor cells and a method thereof, wherein a fluorescence development image of a detected sample is firstly obtained, then the fluorescence development image is subjected to image preprocessing and passes through a shallow feature extractor based on a first convolution neural network model to obtain a fluorescence shallow feature map, then the fluorescence shallow feature map passes through a spatial attention module and a deep feature extractor based on a second convolution neural network model to obtain a fluorescence deep feature map, finally, the spatial enhanced fluorescence shallow feature map and the fluorescence deep feature map are fused, then image semantic segmentation is carried out through a decoder to obtain a segmentation result, and the number of CTC cells is determined based on the segmentation result. In this way, the number of CTC cells is accurately calculated.
In a first aspect, there is provided a fluorescence microscopy imaging system for detection of circulating tumor cells, the system comprising: the data acquisition unit is used for acquiring a fluorescence development image of the detected sample; the image preprocessing unit is used for preprocessing the fluorescence development image to obtain a preprocessed fluorescence development image; the shallow feature extraction unit is used for enabling the preprocessed fluorescence development image to pass through a shallow feature extractor based on a first convolutional neural network model so as to obtain a fluorescence shallow feature map; the space strengthening unit is used for enabling the fluorescence shallow feature map to pass through the space attention module to obtain a space strengthening fluorescence shallow feature map; the deep feature extraction unit is used for enabling the space enhanced fluorescence shallow feature map to pass through a deep feature extractor based on a second convolutional neural network model so as to obtain a fluorescence deep feature map; the feature fusion unit is used for fusing the space enhanced fluorescence shallow feature map and the fluorescence deep feature map to obtain a fluorescence feature map; a decoding unit for passing the fluorescence characteristic map through a decoder to obtain a decoded image; and a counting unit for performing image semantic segmentation on the decoded image to obtain a segmentation result, and determining the number of CTC cells based on the segmentation result.
With reference to the first aspect, in an implementation manner of the first aspect, the shallow feature extraction unit is configured to: input data are respectively carried out in forward transfer of layers by using each layer of the first convolutional neural network model: carrying out convolution processing on the input data to obtain a convolution characteristic diagram; carrying out pooling treatment based on a local feature matrix on the convolution feature map to obtain a pooled feature map; non-linear activation is carried out on the pooled feature map so as to obtain an activated feature map; the input of the first layer of the first convolutional neural network model is the preprocessed fluorescence development image, the output of the last layer of the first convolutional neural network model is the fluorescence shallow layer characteristic diagram, and the number of layers of the first convolutional neural network model is more than or equal to 1 and less than or equal to 6.
With reference to the first aspect, in an implementation manner of the first aspect, the spatial reinforcement unit includes: the pooling unit is used for carrying out average pooling and maximum pooling along the channel dimension on the fluorescence shallow feature map respectively so as to obtain an average feature matrix and a maximum feature matrix; the cascade unit is used for carrying out cascade connection and channel adjustment on the average characteristic matrix and the maximum characteristic matrix to obtain a channel characteristic matrix; the convolution coding unit is used for carrying out convolution coding on the channel characteristic matrix by using a convolution layer of the spatial attention module so as to obtain a convolution characteristic matrix; the score matrix acquisition unit is used for enabling the convolution feature matrix to obtain a space attention score matrix through an activation function; and the attention applying unit is used for multiplying the spatial attention score matrix and each characteristic matrix of the fluorescence shallow characteristic map along the channel dimension by the spatial enhanced fluorescence shallow characteristic map according to position points.
With reference to the first aspect, in an implementation manner of the first aspect, the deep feature extraction unit is configured to: and respectively carrying out convolution processing, pooling processing based on a local feature matrix and nonlinear activation processing on input data in forward transfer of layers by using each layer of the second convolutional neural network model so as to output the fluorescence deep feature map by the last layer of the second convolutional neural network model.
With reference to the first aspect, in an implementation manner of the first aspect, the feature fusion unit is configured to: fusing the spatially enhanced fluorescence shallow feature map and the fluorescence deep feature map using the following fusion formula to obtain a fluorescence feature map; wherein, the fusion formula is:
wherein,,for the fluorescence profile, < >>For the spatially enhanced fluorescence shallow feature map, < >>For the fluorescence deep profile, ">"means that the elements at the corresponding positions of the spatially enhanced fluorescence shallow feature map and the fluorescence deep feature map are added up,">And->Is a weighting parameter for controlling the balance between the spatially enhanced fluorescence shallow feature map and the fluorescence deep feature map in the fluorescence feature map.
With reference to the first aspect, in an implementation manner of the first aspect, the method further includes a training module for training the shallow feature extractor based on the first convolutional neural network model, the spatial attention module, the deep feature extractor based on the second convolutional neural network model, and the decoder; wherein, training module includes: the training data acquisition unit is used for acquiring training data, wherein the training data comprises training fluorescence development images of the detected sample and a true value of the number of CTC cells; the training image preprocessing unit is used for carrying out image preprocessing on the training fluorescence development image to obtain a fluorescence development image after training preprocessing; the training shallow feature extraction unit is used for enabling the fluorescence development image after training pretreatment to pass through the shallow feature extractor based on the first convolutional neural network model so as to obtain a training fluorescence shallow feature map; the training space strengthening unit is used for enabling the training fluorescence shallow feature map to pass through the space attention module so as to obtain a training space strengthening fluorescence shallow feature map; the training deep feature extraction unit is used for enabling the training space enhanced fluorescence shallow feature map to pass through the deep feature extractor based on the second convolutional neural network model so as to obtain a training fluorescence deep feature map; the training feature fusion unit is used for fusing the training space enhanced fluorescence shallow feature map and the training fluorescence deep feature map to obtain a training fluorescence feature map; the optimizing unit is used for optimizing the characteristic manifold curved surface of the training fluorescence characteristic map to obtain an optimized fluorescence characteristic map; the training decoding unit is used for enabling the optimized fluorescence characteristic diagram to pass through the decoder so as to obtain a training decoded image; the training counting unit is used for carrying out image semantic segmentation on the decoded image to obtain a segmentation result, and determining the number of CTC cells based on the segmentation result; and a training unit for calculating a mean square error between the number of CTC cells and a true value of the number of CTC cells as a loss function value, and training the shallow feature extractor based on the first convolutional neural network model, the spatial attention module, the deep feature extractor based on the second convolutional neural network model, and the decoder by a propagation direction of gradient descent.
With reference to the first aspect, in an implementation manner of the first aspect, the optimizing unit is configured to: optimizing the characteristic manifold curved surface of the training fluorescence characteristic map by using the following optimization formula to obtain an optimized fluorescence characteristic map; wherein, the optimization formula is:
wherein the method comprises the steps ofIs the +.f. of the training fluorescence profile>Characteristic value of individual position->And->Is the mean and standard deviation of the feature value sets of each position in the training fluorescence feature map, +.>Represents the maximum function, and->Is the +.f. of the optimized fluorescence profile>Characteristic values of the location.
In a second aspect, a method for monitoring a remote transmission state of a gas pipe network based on data intelligence is provided, and the method comprises the following steps: acquiring a fluorescence development image of a detected sample; performing image preprocessing on the fluorescence development image to obtain a preprocessed fluorescence development image; passing the preprocessed fluorescence development image through a shallow feature extractor based on a first convolutional neural network model to obtain a fluorescence shallow feature map; the fluorescence shallow feature map passes through a spatial attention module to obtain a spatially enhanced fluorescence shallow feature map; the space enhanced fluorescence shallow feature map passes through a deep feature extractor based on a second convolution neural network model to obtain a fluorescence deep feature map; fusing the space enhanced fluorescence shallow layer characteristic map and the fluorescence deep layer characteristic map to obtain a fluorescence characteristic map; passing the fluorescence signature through a decoder to obtain a decoded image; and performing image semantic segmentation on the decoded image to obtain a segmentation result, and determining the number of CTC cells based on the segmentation result.
In a third aspect, there is provided a chip comprising an input-output interface, at least one processor, at least one memory and a bus, the at least one memory to store instructions, the at least one processor to invoke the instructions in the at least one memory to perform the method in the second aspect.
In a fourth aspect, a computer readable medium is provided for storing a computer program comprising instructions for performing the method of the second aspect described above.
In a fifth aspect, there is provided a computer program product comprising instructions which, when executed by a computer, perform the method of the second aspect described above.
According to the fluorescence microscopic imaging system and the method for detecting the circulating tumor cells, firstly, a fluorescence development image of a detected sample is obtained, then, the fluorescence development image is subjected to image preprocessing, a fluorescence shallow feature image is obtained through a shallow feature extractor based on a first convolution neural network model, then, the fluorescence shallow feature image is subjected to a spatial attention module and a deep feature extractor based on a second convolution neural network model to obtain a fluorescence deep feature image, finally, the spatial enhanced fluorescence shallow feature image and the fluorescence deep feature image are fused, then, image semantic segmentation is carried out through a decoder to obtain a segmentation result, and the number of CTC cells is determined based on the segmentation result. In this way, the number of CTC cells is accurately calculated.
Drawings
FIG. 1 is a schematic block diagram of a fluorescence microscopy imaging system for circulating tumor cell detection in accordance with an embodiment of the present application.
FIG. 2 is a schematic diagram of the structure of a training module in a fluorescence microscopy imaging system for circulating tumor cell detection in accordance with an embodiment of the present application.
FIG. 3 is a schematic flow chart of a fluorescence microscopy imaging method for detection of circulating tumor cells according to an embodiment of the present application.
Fig. 4 is a schematic diagram of a model architecture of a fluorescence microscopy imaging method for circulating tumor cell detection in accordance with an embodiment of the present application.
Fig. 5 is a schematic flow chart of a training phase in a fluorescence microscopy imaging method for detection of circulating tumor cells according to an embodiment of the present application.
FIG. 6 is a schematic diagram of a model architecture of a training phase in a fluorescence microscopy imaging method for circulating tumor cell detection in accordance with an embodiment of the present application.
Detailed Description
The technical solutions in the present application will be described below with reference to the accompanying drawings.
Because of the deep learning-based deep neural network model, related terms and concepts of the deep neural network model that may be related to embodiments of the present application are described below.
1. Deep neural network model
In the deep neural network model, the hidden layers may be convolutional layers and pooled layers. The set of weight values corresponding to the convolutional layer is referred to as a filter, also referred to as a convolutional kernel. The filter and the input eigenvalue are both represented as a multi-dimensional matrix, correspondingly, the filter represented as a multi-dimensional matrix is also called a filter matrix, the input eigenvalue represented as a multi-dimensional matrix is also called an input eigenvalue, of course, besides the input eigenvalue, the eigenvector can also be input, and the input eigenvector is only exemplified by the input eigenvector. The operation of the convolution layer is called a convolution operation, which is to perform an inner product operation on a part of eigenvalues of the input eigenvalue matrix and weight values of the filter matrix.
The operation process of each convolution layer in the deep neural network model can be programmed into software, and then the output result of each layer of network, namely the output characteristic matrix, is obtained by running the software in an operation device. For example, the software performs inner product operation by taking the upper left corner of the input feature matrix of each layer of network as a starting point and taking the size of the filter as a window in a sliding window mode, and extracting data of one window from the feature value matrix each time. After the inner product operation is completed between the data of the right lower corner window of the input feature matrix and the filter, a two-dimensional output feature matrix of each layer of network can be obtained. The software repeats the above process until the entire output feature matrix for each layer of network is generated.
The convolution layer operation process is to slide a window with a filter size across the whole input image (i.e. the input feature matrix), and at each moment, to perform inner product operation on the input feature value covered in the window and the filter, wherein the step length of window sliding is 1. Specifically, the upper left corner of the input feature matrix is used as a starting point, the size of the filter is used as a window, the sliding step length of the window is 1, the input feature value of one window is extracted from the feature value matrix each time and the filter performs inner product operation, and when the data of the lower right corner of the input feature matrix and the filter complete inner product operation, a two-dimensional output feature matrix of the input feature matrix can be obtained.
Since it is often necessary to reduce the number of training parameters, the convolutional layer often requires a periodic introduction of a pooling layer, the only purpose of which is to reduce the spatial size of the image during image processing. The pooling layer may include an average pooling operator and/or a maximum pooling operator for sampling the input image to obtain a smaller size image. The average pooling operator may calculate pixel values in the image over a particular range to produce an average as a result of the average pooling. The max pooling operator may take the pixel with the largest value in a particular range as the result of max pooling. In addition, just as the size of the weighting matrix used in the convolutional layer should be related to the image size, the operators in the pooling layer should also be related to the image size. The size of the image output after the processing by the pooling layer can be smaller than the size of the image input to the pooling layer, and each pixel point in the image output by the pooling layer represents the average value or the maximum value of the corresponding sub-region of the image input to the pooling layer.
Since the functions actually required to be simulated in the deep neural network are nonlinear, but the previous rolling and pooling can only simulate linear functions, in order to introduce nonlinear factors in the deep neural network model to increase the characterization capacity of the whole network, an activation layer is further arranged after the pooling layer, an activation function is arranged in the activation layer, and the commonly used excitation functions include sigmoid, tanh, reLU functions and the like.
2. Softmax classification function
The Softmax classification function is also called soft maximum function, normalized exponential function. One K-dimensional vector containing arbitrary real numbers can be "compressed" into another K-dimensional real vector such that each element ranges between (0, 1) and the sum of all elements is 1. The Softmax classification function is commonly used to classify problems.
Having described the relevant terms and concepts of the deep neural network model that may be involved in embodiments of the present application, the following description will be made of the basic principles of the present application for the convenience of understanding by those skilled in the art.
3. Attention (attention) mechanism
In colloquial terms, the attention mechanism is to focus attention on important points, while ignoring other unimportant factors. For example, the attention mechanism is similar to the visual attention mechanism of human, and human vision can obtain a target area needing to be focused on, namely a focus of attention, by rapidly scanning a global image when facing the image, and then throw more attention resources into the area to acquire more detail information needing to be focused on, and inhibit other useless information. The determination of the importance level may depend on the application scenario, among other things.
In view of the above technical problems, the technical idea of the present application is to acquire a fluorescence developed image of a sample to be detected, to divide CTC cells from the fluorescence developed image using an image processing technique based on deep learning, and to accurately calculate the number of CTC cells.
Specifically, in the technical scheme of the present application, first, a fluorescence developed image of a sample to be detected is acquired. Here, the fluorescence developed image is an image in which a sample to be detected is irradiated with ultraviolet rays by using the principle of a fluorescence microscope to emit fluorescence, and then the shape and position thereof are observed under the microscope. The effective information and basis can be provided for subsequent CTC detection and analysis by acquiring a fluorescence development image of the detected sample.
Then, the fluorescence developed image is subjected to image preprocessing to obtain a preprocessed fluorescence developed image. Here, the use of image preprocessing can suppress unwanted distortion or enhance important image features, and specifically, can be performed using operations such as noise removal, contrast enhancement, brightness adjustment, and the like. In general, image preprocessing can improve the quality of images, reduce the information quantity, reduce the computational complexity and improve the accuracy and efficiency of subsequent feature extraction and segmentation.
Considering that the edge, texture and shape information in the preprocessed fluorescence development image have important significance for distinguishing CTC cells from normal blood cells, in the technical scheme of the application, the preprocessed fluorescence development image is passed through a shallow feature extractor based on a first convolutional neural network model to obtain a fluorescence shallow feature map. That is, the shallow feature extractor based on the first convolutional neural network model is utilized to extract local shallow feature information in the preprocessed fluoroscopic image.
And then, the fluorescence shallow feature map passes through a spatial attention module to obtain a spatially enhanced fluorescence shallow feature map. Here, the spatial attention module may dynamically adjust the weights of the spatial locations such that useful information is enhanced while irrelevant information is suppressed. Specifically, the spatial attention module can be used for strengthening the spatial characteristics of the data by weighting different parts of the input data. That is, the fluorescence shallow feature map is input into the spatial attention module, and the module automatically selects and intensifies the feature parts having discrimination on CTC cells and suppresses the feature parts irrelevant to classification, thereby obtaining the spatially intensified fluorescence shallow feature map. The fluorescence shallow feature map after spatial enhancement can better reflect the difference and relationship between CTC cells and other cells.
Further, the space enhanced fluorescence shallow feature map is passed through a deep feature extractor based on a second convolutional neural network model to obtain a fluorescence deep feature map. The deep feature extractor based on the second convolutional neural network model performs further abstraction on the space enhanced fluorescence shallow feature map through a plurality of convolutional layers and a pooling layer, so that semantic information is enhanced.
In order to keep the detail information of the shallow features and the semantic information of the deep features, in the technical scheme of the application, the space enhanced fluorescence shallow feature map and the fluorescence deep feature map are fused to obtain a fluorescence feature map. Thus, the fluorescence characteristic diagram has more abundant characteristic expression capability.
The fluorescence signature is then passed through a decoder to obtain a decoded image. Specifically, the decoder can decode the characteristic information of the position and morphology of CTC cells in the regression fluorescence characteristic map by deconvolution. Subsequently, image semantic segmentation is performed on the decoded image to obtain a segmentation result, and the number of CTC cells is determined based on the segmentation result. Wherein the image semantic segmentation is capable of learning features of the image and predicting a class of each pixel from the features. In this way CTC cells can be distinguished from other cells or background.
Here, when the spatially enhanced fluorescence shallow feature map and the fluorescence deep feature map are fused to obtain a fluorescence feature map, in order to fully utilize the image semantic shallow feature and the deep feature expressed by the spatially enhanced fluorescence shallow feature map and the fluorescence deep feature map, the fluorescence feature map is preferably obtained by directly cascading the spatially enhanced fluorescence shallow feature map and the fluorescence deep feature map along a channel dimension, so that when the fluorescence feature map is decoded, since the fluorescence feature map simultaneously contains the image semantic shallow feature and the deep feature with consistency differences, the convergence effect of the fluorescence feature map through a decoder is poor due to the consistency differences between the feature values of the fluorescence feature map, and the training speed of a model is reduced.
Thus, in the context of the present application, the fluorescence profile is, for example, denoted asCarrying out tangential plane directional distance normalization of the characteristic manifold curved surface based on the neighborhood points, wherein the tangential plane directional distance normalization concretely comprises the following steps:
wherein the method comprises the steps ofAnd->Is a feature value set +.>Mean and standard deviation of (2), and->Is the +.f. of the optimized classification feature graph>Characteristic values of the location.
Here, neighborhood point-based tangent plane directed distance normalization of the feature manifold surface may be performed by the fluorescence feature map To construct a local linear tangent space based on a statistical neighborhood for each feature value thereof, to orthographically orient feature values by selecting a maximum geometric measure of tangent vectors within the local linear tangent space, and to normalize the expression of local non-European geometric properties for points on a flow surface based on an inner product distance expression of the orientation vectors, to thereby promote the fluorescent feature map by means of geometric correction of the flow surface>The expression consistency of each characteristic value of the high-dimensional characteristic set is improved, so that the convergence effect of the optimized fluorescent characteristic diagram through a decoder is poor, and the training speed of the model is improved.
The application has the following technical effects:
1. a fluorescence microscopy imaging protocol for the detection of circulating tumor cells is provided, and more particularly, an intelligent detection and counting protocol for circulating tumor cells.
2. According to the scheme, the CTC cells can be effectively separated from the fluorescence development image, so that the CTC cells and other blood cells are distinguished, and false detection or omission caused by similar shapes and sizes is avoided. Meanwhile, the number of CTC cells can be accurately calculated, and valuable information is provided for early diagnosis and treatment of tumors.
Having described the basic principles of the present application, various non-limiting embodiments of the present application are described in detail below with reference to the accompanying drawings.
FIG. 1 is a schematic block diagram of a fluorescence microscopy imaging system for circulating tumor cell detection in accordance with an embodiment of the present application. As shown in fig. 1, a fluorescence microscopy imaging system 100 for detection of circulating tumor cells, comprising:
a data acquisition unit 110 for acquiring a fluorescence developed image of the sample to be detected. It should be understood that a fluorescence developed image is an image of the shape and position of a sample to be inspected by irradiating it with ultraviolet rays using the principle of a fluorescence microscope, causing it to fluoresce, and then observing it under the microscope. The effective information and basis can be provided for subsequent CTC detection and analysis by acquiring a fluorescence development image of the detected sample.
The image preprocessing unit 120 is configured to perform image preprocessing on the fluorescence developed image to obtain a preprocessed fluorescence developed image. It will be appreciated that the use of image pre-processing may suppress unwanted distortion or enhance important image features, and in particular may be done using operations such as noise removal, contrast enhancement, brightness adjustment, etc. In general, image preprocessing can improve the quality of images, reduce the information quantity, reduce the computational complexity and improve the accuracy and efficiency of subsequent feature extraction and segmentation.
And the shallow feature extraction unit 130 is used for passing the preprocessed fluorescence development image through a shallow feature extractor based on a first convolutional neural network model to obtain a fluorescence shallow feature map. It should be understood that, considering that the edge, texture and shape information in the pre-processed fluorescence developed image is of great importance for distinguishing CTC cells from normal blood cells, in the technical solution of the present application, the pre-processed fluorescence developed image is passed through a shallow feature extractor based on a first convolutional neural network model to obtain a fluorescence shallow feature map. That is, the shallow feature extractor based on the first convolutional neural network model is utilized to extract local shallow feature information in the preprocessed fluoroscopic image.
Optionally, in an embodiment of the present application, the shallow feature extraction unit is configured to: input data are respectively carried out in forward transfer of layers by using each layer of the first convolutional neural network model: carrying out convolution processing on the input data to obtain a convolution characteristic diagram; carrying out pooling treatment based on a local feature matrix on the convolution feature map to obtain a pooled feature map; non-linear activation is carried out on the pooled feature map so as to obtain an activated feature map; the input of the first layer of the first convolutional neural network model is the preprocessed fluorescence development image, the output of the last layer of the first convolutional neural network model is the fluorescence shallow layer characteristic diagram, and the number of layers of the first convolutional neural network model is more than or equal to 1 and less than or equal to 6.
The space enhancement unit 140 is configured to pass the fluorescence shallow feature map through a space attention module to obtain a space enhanced fluorescence shallow feature map. It should be appreciated that the spatial attention module may dynamically adjust the weights of the spatial locations such that useful information is enhanced while irrelevant information is suppressed. Specifically, the spatial attention module can be used for strengthening the spatial characteristics of the data by weighting different parts of the input data. That is, the fluorescence shallow feature map is input into the spatial attention module, and the module automatically selects and intensifies the feature parts having discrimination on CTC cells and suppresses the feature parts irrelevant to classification, thereby obtaining the spatially intensified fluorescence shallow feature map. The fluorescence shallow feature map after spatial enhancement can better reflect the difference and relationship between CTC cells and other cells.
Optionally, in an embodiment of the present application, the space enhancement unit includes: the pooling unit is used for carrying out average pooling and maximum pooling along the channel dimension on the fluorescence shallow feature map respectively so as to obtain an average feature matrix and a maximum feature matrix; the cascade unit is used for carrying out cascade connection and channel adjustment on the average characteristic matrix and the maximum characteristic matrix to obtain a channel characteristic matrix; the convolution coding unit is used for carrying out convolution coding on the channel characteristic matrix by using a convolution layer of the spatial attention module so as to obtain a convolution characteristic matrix; the score matrix acquisition unit is used for enabling the convolution feature matrix to obtain a space attention score matrix through an activation function; and the attention applying unit is used for multiplying the spatial attention score matrix and each characteristic matrix of the fluorescence shallow characteristic map along the channel dimension by the spatial enhanced fluorescence shallow characteristic map according to position points.
And a deep feature extraction unit 150, configured to pass the spatially enhanced fluorescence shallow feature map through a deep feature extractor based on a second convolutional neural network model to obtain a fluorescence deep feature map. It should be appreciated that considering that the neural network model is used in feature extraction, if only shallow features are extracted, the accuracy is not high, but deep features can be extracted without doubt to enhance the accuracy of the final result. Thus, the spatially enhanced fluorescence shallow feature map is passed through a deep feature extractor based on a second convolutional neural network model to obtain a fluorescence deep feature map. The deep feature extractor based on the second convolutional neural network model performs further abstraction on the space enhanced fluorescence shallow feature map through a plurality of convolutional layers and a pooling layer, so that semantic information is enhanced.
Optionally, in an embodiment of the present application, the deep feature extraction unit is configured to: and respectively carrying out convolution processing, pooling processing based on a local feature matrix and nonlinear activation processing on input data in forward transfer of layers by using each layer of the second convolutional neural network model so as to output the fluorescence deep feature map by the last layer of the second convolutional neural network model. Here, a ratio of the number of layers of the second convolutional neural network model to the number of layers of the first convolutional neural network model is 5 or more.
And a feature fusion unit 160, configured to fuse the spatially enhanced fluorescence shallow feature map and the fluorescence deep feature map to obtain a fluorescence feature map. It should be understood that, in order to preserve the detail information of the shallow features and the semantic information of the deep features, in the technical solution of the present application, the spatially enhanced fluorescence shallow feature map and the fluorescence deep feature map are fused to obtain a fluorescence feature map. Thus, the fluorescence characteristic diagram has more abundant characteristic expression capability.
Optionally, in an embodiment of the present application, the feature fusion unit is configured to: fusing the spatially enhanced fluorescence shallow feature map and the fluorescence deep feature map using the following fusion formula to obtain a fluorescence feature map; wherein, the fusion formula is:
wherein,,for the fluorescence profile, < >>For the spatially enhanced fluorescence shallow feature map, < >>For the fluorescence deep profile, ">"means that the elements at the corresponding positions of the spatially enhanced fluorescence shallow feature map and the fluorescence deep feature map are added up,">And->Is a weighting parameter for controlling the balance between the spatially enhanced fluorescence shallow feature map and the fluorescence deep feature map in the fluorescence feature map.
And a decoding unit 170 for passing the fluorescence characteristic map through a decoder to obtain a decoded image. It will be appreciated that the characteristic information of the position, morphology of CTC cells in the regression fluorescence profile can be decoded by deconvolution using a decoder.
And a counting unit 180 for performing image semantic segmentation on the decoded image to obtain a segmentation result, and determining the number of CTC cells based on the segmentation result. It should be appreciated that the decoded image is subjected to image semantic segmentation to obtain segmentation results, and the number of CTC cells is determined based on the segmentation results. Wherein the image semantic segmentation is capable of learning features of the image and predicting a class of each pixel from the features. In this way CTC cells can be distinguished from other cells or background. Here, one skilled in the art will appreciate that an image is made up of many pixels, and semantic segmentation, as the name implies, is the grouping or segmentation of pixels according to differences in the meaning of the semantics of the representation of the image.
FIG. 2 is a schematic block diagram of a training module in a fluorescence microscopy imaging system for circulating tumor cell detection in accordance with an embodiment of the present application. As shown in fig. 2, in an embodiment of the present application, the fluorescence microscopy imaging system for detecting circulating tumor cells further includes a training module 200 for training the shallow feature extractor based on the first convolutional neural network model, the spatial attention module, the deep feature extractor based on the second convolutional neural network model, and the decoder; wherein the training module 200 comprises: a training data acquisition unit 210 for acquiring training data including a training fluorescence developed image of the sample to be detected and a true value of the number of CTC cells; a training image preprocessing unit 220, configured to perform image preprocessing on the training fluorescence development image to obtain a fluorescence development image after training preprocessing; a training shallow feature extraction unit 230, configured to pass the fluorescence development image after training pretreatment through the shallow feature extractor based on the first convolutional neural network model to obtain a training fluorescence shallow feature map; a training space enhancement unit 240, configured to pass the training fluorescence shallow feature map through the spatial attention module to obtain a training space enhanced fluorescence shallow feature map; a training deep feature extraction unit 250, configured to pass the training space enhanced fluorescence shallow feature map through the deep feature extractor based on the second convolutional neural network model to obtain a training fluorescence deep feature map; a training feature fusion unit 260, configured to fuse the training space enhanced fluorescence shallow feature map and the training fluorescence deep feature map to obtain a training fluorescence feature map; an optimizing unit 270, configured to perform feature manifold surface optimization on the training fluorescence feature map to obtain an optimized fluorescence feature map; a training decoding unit 280, configured to pass the optimized fluorescence feature map through the decoder to obtain a training decoded image; a training counting unit 290 for performing image semantic segmentation on the decoded image to obtain a segmentation result, and determining the number of CTC cells based on the segmentation result; and a training unit 300 for calculating a mean square error between the number of CTC cells and a true value of the number of CTC cells as a loss function value, and training the shallow feature extractor based on the first convolutional neural network model, the spatial attention module, the deep feature extractor based on the second convolutional neural network model, and the decoder by a propagation direction of gradient descent.
When the space enhanced fluorescence shallow feature map and the fluorescence deep feature map are fused to obtain a fluorescence feature map, in order to fully utilize the image semantic shallow feature and the deep feature expressed by the space enhanced fluorescence shallow feature map and the fluorescence deep feature map, the fluorescence feature map is preferably obtained by directly cascading the space enhanced fluorescence shallow feature map and the fluorescence deep feature map along a channel dimension, so that when the fluorescence feature map is decoded, as the fluorescence feature map simultaneously contains the image semantic shallow feature and the deep feature with consistency difference, the convergence effect of the fluorescence feature map through a decoder is poor due to the consistency difference between feature values of the fluorescence feature map, and the training speed of a model is reduced. Thus, in the context of the present application, the fluorescence profile is, for example, denoted asAnd carrying out tangential plane directional distance normalization of the characteristic manifold curved surface based on the neighborhood points.
Optionally, in an embodiment of the present application, the optimizing unit is configured to: optimizing the characteristic manifold curved surface of the training fluorescence characteristic map by using the following optimization formula to obtain an optimized fluorescence characteristic map;
Wherein, the optimization formula is:
wherein the method comprises the steps ofIs the +.f. of the training fluorescence profile>Characteristic value of individual position->And->Is the mean and standard deviation of the feature value sets of each position in the training fluorescence feature map, +.>Represents the maximum function, and->Is the +.f. of the optimized fluorescence profile>Characteristic values of the location.
Here, neighborhood point-based tangent plane directed distance normalization of the feature manifold surface may be performed by the fluorescence feature mapTo construct a statistical neighborhood based local linear tangent space for each feature value thereof, to orthographically orient feature values by selecting a maximum geometric measure of tangent vectors within the local linear tangent space, and to orthographically express local non-euro-geometric properties of points on a surface of a flow based on an inner product distance expression of orientation vectorsThereby improving the fluorescence profile by means of geometric correction of the manifold curve>The expression consistency of each characteristic value of the high-dimensional characteristic set is improved, so that the convergence effect of the optimized fluorescent characteristic diagram through a decoder is poor, and the training speed of the model is improved.
In summary, the fluorescence microscopic imaging system for detecting the circulating tumor cells firstly obtains a fluorescence development image of a detected sample, then performs image preprocessing on the fluorescence development image, obtains a fluorescence shallow feature map through a shallow feature extractor based on a first convolutional neural network model, then obtains a fluorescence deep feature map through a spatial attention module and a deep feature extractor based on a second convolutional neural network model, and finally fuses the spatially enhanced fluorescence shallow feature map and the fluorescence deep feature map, performs image semantic segmentation through a decoder to obtain a segmentation result, and determines the number of CTC cells based on the segmentation result. In this way, the number of CTC cells is accurately calculated.
FIG. 3 is a schematic flow chart of a fluorescence microscopy imaging method for detection of circulating tumor cells according to an embodiment of the present application. As shown in fig. 3, the fluorescence microscopic imaging method for detecting the circulating tumor cells comprises the following steps: s110, acquiring a fluorescence development image of a detected sample; s120, carrying out image preprocessing on the fluorescence development image to obtain a preprocessed fluorescence development image; s130, passing the preprocessed fluorescence development image through a shallow feature extractor based on a first convolutional neural network model to obtain a fluorescence shallow feature map; s140, the fluorescence shallow feature map passes through a spatial attention module to obtain a spatially enhanced fluorescence shallow feature map; s150, passing the space enhanced fluorescence shallow feature map through a deep feature extractor based on a second convolutional neural network model to obtain a fluorescence deep feature map; s160, fusing the space enhanced fluorescence shallow layer characteristic map and the fluorescence deep layer characteristic map to obtain a fluorescence characteristic map; s170, passing the fluorescence characteristic map through a decoder to obtain a decoded image; and S180, performing image semantic segmentation on the decoded image to obtain a segmentation result, and determining the number of CTC cells based on the segmentation result.
Fig. 4 is a schematic diagram of a model architecture of a fluorescence microscopy imaging method for circulating tumor cell detection in accordance with an embodiment of the present application. As shown in fig. 4, the input of the model architecture of the fluorescence microscopic imaging method for detecting the circulating tumor cells is a fluorescence development image of the detected sample. Firstly, carrying out image preprocessing on the fluorescence developed image to obtain a preprocessed fluorescence developed image, and enabling the preprocessed fluorescence developed image to pass through a shallow feature extractor based on a first convolution neural network model to obtain a fluorescence shallow feature map. And then, the fluorescence shallow feature map passes through a spatial attention module to obtain a spatially enhanced fluorescence shallow feature map, and the spatially enhanced fluorescence shallow feature map passes through a deep feature extractor based on a second convolutional neural network model to obtain a fluorescence deep feature map. And then, fusing the space enhanced fluorescence shallow layer characteristic map and the fluorescence deep layer characteristic map to obtain a fluorescence characteristic map, and enabling the fluorescence characteristic map to pass through a decoder to obtain a decoded image. Finally, performing image semantic segmentation on the decoded image to obtain a segmentation result, and determining the number of CTC cells based on the segmentation result
Fig. 5 is a schematic flow chart of a training phase in a fluorescence microscopy imaging method for detection of circulating tumor cells according to an embodiment of the present application. As shown in fig. 5, the method further includes a training phase for training the shallow feature extractor based on the first convolutional neural network model, the spatial attention module, the deep feature extractor based on the second convolutional neural network model, and the decoder; wherein the training phase comprises: s210, acquiring training data, wherein the training data comprises training fluorescence development images of detected samples and true values of the number of CTC cells; s220, performing image preprocessing on the training fluorescence development image to obtain a fluorescence development image after training preprocessing; s230, passing the fluorescence development image after training pretreatment through the shallow feature extractor based on the first convolutional neural network model to obtain a training fluorescence shallow feature map; s240, passing the training fluorescence shallow feature map through the spatial attention module to obtain a training spatial enhanced fluorescence shallow feature map; s250, passing the training space enhanced fluorescence shallow feature map through the deep feature extractor based on the second convolutional neural network model to obtain a training fluorescence deep feature map; s260, fusing the training space enhanced fluorescence shallow feature map and the training fluorescence deep feature map to obtain a training fluorescence feature map; s270, optimizing the characteristic manifold curved surface of the training fluorescence characteristic map to obtain an optimized fluorescence characteristic map; s280, enabling the optimized fluorescence characteristic diagram to pass through the decoder to obtain a training decoded image; s290, performing image semantic segmentation on the decoded image to obtain a segmentation result, and determining the number of CTC cells based on the segmentation result; and S300, calculating the mean square error between the number of the CTC cells and the true value of the number of the CTC cells as a loss function value, and training the shallow feature extractor based on the first convolutional neural network model, the spatial attention module, the deep feature extractor based on the second convolutional neural network model and the decoder through the propagation direction of gradient descent.
FIG. 6 is a schematic diagram of a model architecture of a training phase in a fluorescence microscopy imaging method for circulating tumor cell detection in accordance with an embodiment of the present application. As shown in fig. 6, the input of the model architecture of the training phase is a training fluoroscopic image of the sample being tested. Firstly, carrying out image preprocessing on the training fluorescence development image to obtain a training preprocessed fluorescence development image, and enabling the training preprocessed fluorescence development image to pass through the shallow feature extractor based on the first convolutional neural network model to obtain a training fluorescence shallow feature map. And then, the training fluorescence shallow feature map passes through the spatial attention module to obtain a training spatial enhanced fluorescence shallow feature map, and the training spatial enhanced fluorescence shallow feature map passes through the deep feature extractor based on the second convolutional neural network model to obtain a training fluorescence deep feature map. And then, fusing the training space enhanced fluorescence shallow feature map and the training fluorescence deep feature map to obtain a training fluorescence feature map, and carrying out feature manifold curved surface optimization on the training fluorescence feature map to obtain an optimized fluorescence feature map. And then, the optimized fluorescence characteristic map passes through the decoder to obtain a training decoded image, image semantic segmentation is carried out on the decoded image to obtain a segmentation result, and the number of CTC cells is determined based on the segmentation result. Finally, calculating the mean square error between the number of CTC cells and the true value of the number of CTC cells as a loss function value, and training the shallow feature extractor based on the first convolutional neural network model, the spatial attention module, the deep feature extractor based on the second convolutional neural network model and the decoder through the propagation direction of gradient descent.
Optionally, in an embodiment of the present application, performing feature manifold surface optimization on the training fluorescence feature map to obtain an optimized fluorescence feature map includes: optimizing the characteristic manifold curved surface of the training fluorescence characteristic map by using the following optimization formula to obtain an optimized fluorescence characteristic map;
wherein, the optimization formula is:
wherein the method comprises the steps ofIs the +.f. of the training fluorescence profile>Characteristic value of individual position->And->Is the mean and standard deviation of the feature value sets of each position in the training fluorescence feature map, +.>Represents the maximum function, and->Is the +.f. of the optimized fluorescence profile>Characteristic values of the location.
Here, it will be understood by those skilled in the art that the specific operations of the respective steps in the above-described fluorescence microscopy imaging method for detection of circulating tumor cells have been described in detail in the above description of the fluorescence microscopy imaging system for detection of circulating tumor cells with reference to fig. 1 to 2, and thus, repetitive descriptions thereof will be omitted.
The embodiment of the invention also provides a chip system, which comprises at least one processor, and when the program instructions are executed in the at least one processor, the method provided by the embodiment of the application is realized.
The embodiment of the invention also provides a computer storage medium, on which a computer program is stored, which when executed by a computer causes the computer to perform the method of the above-described method embodiment.
The present invention also provides a computer program product comprising instructions which, when executed by a computer, cause the computer to perform the method of the method embodiment described above.
In the above embodiments, it may be implemented in whole or in part by software, hardware, firmware, or any other combination. When implemented in software, may be implemented in whole or in part in the form of a computer program product. The computer program product includes one or more computer instructions. When loaded and executed on a computer, produces a flow or function in accordance with embodiments of the present invention, in whole or in part. The computer may be a general purpose computer, a special purpose computer, a computer network, or other programmable apparatus. The computer instructions may be stored in a computer-readable storage medium or transmitted from one computer-readable storage medium to another computer-readable storage medium, for example, the computer instructions may be transmitted from one website, computer, server, or data center to another website, computer, server, or data center by a wired (e.g., coaxial cable, fiber optic, digital subscriber line (digital subscriber line, DSL)) or wireless (e.g., infrared, wireless, microwave, etc.). The computer readable storage medium may be any available medium that can be accessed by a computer or a data storage device such as a server, data center, etc. that contains an integration of one or more available media. The usable medium may be a magnetic medium (e.g., a floppy disk, a hard disk, a magnetic tape), an optical medium (e.g., a digital video disc (digital video disc, DVD)), or a semiconductor medium (e.g., a Solid State Disk (SSD)), or the like.
Those of ordinary skill in the art will appreciate that the various illustrative elements and algorithm steps described in connection with the embodiments disclosed herein may be implemented as electronic hardware, or combinations of computer software and electronic hardware. Whether such functionality is implemented as hardware or software depends upon the particular application and design constraints imposed on the solution. Skilled artisans may implement the described functionality in varying ways for each particular application, but such implementation decisions should not be interpreted as causing a departure from the scope of the present invention.
In the several embodiments provided by the present invention, it should be understood that the disclosed systems, devices, and methods may be implemented in other manners. For example, the apparatus embodiments described above are merely illustrative, e.g., the division of the units is merely a logical function division, and there may be additional divisions when actually implemented, e.g., multiple units or components may be combined or integrated into another system, or some features may be omitted or not performed. Alternatively, the coupling or direct coupling or communication connection shown or discussed with each other may be an indirect coupling or communication connection via some interfaces, devices or units, which may be in electrical, mechanical or other form.
The units described as separate units may or may not be physically separate, and units shown as units may or may not be physical units, may be located in one place, or may be distributed on a plurality of network units. Some or all of the units may be selected according to actual needs to achieve the purpose of the solution of this embodiment.
In addition, each functional unit in the embodiments of the present invention may be integrated in one processing unit, or each unit may exist alone physically, or two or more units may be integrated in one unit.
Claims (10)
1. A fluorescence microscopy imaging system for detection of circulating tumor cells, comprising:
the data acquisition unit is used for acquiring a fluorescence development image of the detected sample;
the image preprocessing unit is used for preprocessing the fluorescence development image to obtain a preprocessed fluorescence development image;
the shallow feature extraction unit is used for enabling the preprocessed fluorescence development image to pass through a shallow feature extractor based on a first convolutional neural network model so as to obtain a fluorescence shallow feature map;
the space strengthening unit is used for enabling the fluorescence shallow feature map to pass through the space attention module to obtain a space strengthening fluorescence shallow feature map;
The deep feature extraction unit is used for enabling the space enhanced fluorescence shallow feature map to pass through a deep feature extractor based on a second convolutional neural network model so as to obtain a fluorescence deep feature map;
the feature fusion unit is used for fusing the space enhanced fluorescence shallow feature map and the fluorescence deep feature map to obtain a fluorescence feature map;
a decoding unit for passing the fluorescence characteristic map through a decoder to obtain a decoded image; and
and the counting unit is used for carrying out image semantic segmentation on the decoded image to obtain a segmentation result and determining the number of CTC cells based on the segmentation result.
2. The fluorescence microscopy imaging system for detection of circulating tumor cells of claim 1, wherein the shallow feature extraction unit is configured to: input data are respectively carried out in forward transfer of layers by using each layer of the first convolutional neural network model:
carrying out convolution processing on the input data to obtain a convolution characteristic diagram;
carrying out pooling treatment based on a local feature matrix on the convolution feature map to obtain a pooled feature map;
non-linear activation is carried out on the pooled feature map so as to obtain an activated feature map;
The input of the first layer of the first convolutional neural network model is the preprocessed fluorescence development image, the output of the last layer of the first convolutional neural network model is the fluorescence shallow layer characteristic diagram, and the number of layers of the first convolutional neural network model is more than or equal to 1 and less than or equal to 6.
3. The fluorescence microscopy imaging system for detection of circulating tumor cells of claim 2, wherein the spatial enhancement unit comprises:
the pooling unit is used for carrying out average pooling and maximum pooling along the channel dimension on the fluorescence shallow feature map respectively so as to obtain an average feature matrix and a maximum feature matrix;
the cascade unit is used for carrying out cascade connection and channel adjustment on the average characteristic matrix and the maximum characteristic matrix to obtain a channel characteristic matrix;
the convolution coding unit is used for carrying out convolution coding on the channel characteristic matrix by using a convolution layer of the spatial attention module so as to obtain a convolution characteristic matrix;
the score matrix acquisition unit is used for enabling the convolution feature matrix to obtain a space attention score matrix through an activation function;
and the attention applying unit is used for multiplying the spatial attention score matrix and each characteristic matrix of the fluorescence shallow characteristic map along the channel dimension by the spatial enhanced fluorescence shallow characteristic map according to position points.
4. The fluorescence microscopy imaging system for detection of circulating tumor cells of claim 3, wherein the deep feature extraction unit is configured to: and respectively carrying out convolution processing, pooling processing based on a local feature matrix and nonlinear activation processing on input data in forward transfer of layers by using each layer of the second convolutional neural network model so as to output the fluorescence deep feature map by the last layer of the second convolutional neural network model.
5. The fluorescence microscopy imaging system for detection of circulating tumor cells of claim 4, wherein the feature fusion unit is configured to: fusing the spatially enhanced fluorescence shallow feature map and the fluorescence deep feature map using the following fusion formula to obtain the fluorescence feature map;
wherein, the fusion formula is:
wherein,,for the fluorescence profile, < >>For the spatially enhanced fluorescence shallow feature map, < >>For the fluorescence deep profile, ">"means that the elements at the corresponding positions of the spatially enhanced fluorescence shallow feature map and the fluorescence deep feature map are added up,">And->Is a weighting parameter for controlling the balance between the spatially enhanced fluorescence shallow feature map and the fluorescence deep feature map in the fluorescence feature map.
6. The fluorescence microscopy imaging system of claim 5, further comprising a training module for training the shallow feature extractor based on the first convolutional neural network model, the spatial attention module, the deep feature extractor based on the second convolutional neural network model, and the decoder;
wherein, training module includes:
the training data acquisition unit is used for acquiring training data, wherein the training data comprises training fluorescence development images of the detected sample and a true value of the number of CTC cells;
the training image preprocessing unit is used for carrying out image preprocessing on the training fluorescence development image to obtain a fluorescence development image after training preprocessing;
the training shallow feature extraction unit is used for enabling the fluorescence development image after training pretreatment to pass through the shallow feature extractor based on the first convolutional neural network model so as to obtain a training fluorescence shallow feature map;
the training space strengthening unit is used for enabling the training fluorescence shallow feature map to pass through the space attention module so as to obtain a training space strengthening fluorescence shallow feature map;
The training deep feature extraction unit is used for enabling the training space enhanced fluorescence shallow feature map to pass through the deep feature extractor based on the second convolutional neural network model so as to obtain a training fluorescence deep feature map;
the training feature fusion unit is used for fusing the training space enhanced fluorescence shallow feature map and the training fluorescence deep feature map to obtain a training fluorescence feature map;
the optimizing unit is used for optimizing the characteristic manifold curved surface of the training fluorescence characteristic map to obtain an optimized fluorescence characteristic map;
the training decoding unit is used for enabling the optimized fluorescence characteristic diagram to pass through the decoder so as to obtain a training decoded image;
the training counting unit is used for carrying out image semantic segmentation on the decoded image to obtain a segmentation result, and determining the number of CTC cells based on the segmentation result; and
and the training unit is used for calculating the mean square error between the number of the CTC cells and the true value of the number of the CTC cells as a loss function value and training the shallow feature extractor based on the first convolutional neural network model, the spatial attention module, the deep feature extractor based on the second convolutional neural network model and the decoder through the propagation direction of gradient descent.
7. The fluorescence microscopy imaging system for detection of circulating tumor cells of claim 6, wherein the optimization unit is configured to: optimizing the characteristic manifold curved surface of the training fluorescence characteristic map by using the following optimization formula to obtain the optimized fluorescence characteristic map;
wherein, the optimization formula is:
wherein the method comprises the steps ofIs the +.f. of the training fluorescence profile>Characteristic value of individual position->And->Is the mean and standard deviation of the feature value sets of each position in the training fluorescence feature map, +.>Represents the maximum function, and->Is the +.f. of the optimized fluorescence profile>Characteristic values of the location.
8. A fluorescence microscopy imaging method for detecting circulating tumor cells, comprising:
acquiring a fluorescence development image of a detected sample;
performing image preprocessing on the fluorescence development image to obtain a preprocessed fluorescence development image;
passing the preprocessed fluorescence development image through a shallow feature extractor based on a first convolutional neural network model to obtain a fluorescence shallow feature map;
the fluorescence shallow feature map passes through a spatial attention module to obtain a spatially enhanced fluorescence shallow feature map;
the space enhanced fluorescence shallow feature map passes through a deep feature extractor based on a second convolution neural network model to obtain a fluorescence deep feature map;
Fusing the space enhanced fluorescence shallow layer characteristic map and the fluorescence deep layer characteristic map to obtain a fluorescence characteristic map;
passing the fluorescence signature through a decoder to obtain a decoded image; and
and performing image semantic segmentation on the decoded image to obtain a segmentation result, and determining the number of CTC cells based on the segmentation result.
9. The fluorescence microscopy imaging method for circulating tumor cell detection of claim 8, further comprising a training phase for training the shallow feature extractor based on the first convolutional neural network model, the spatial attention module, the deep feature extractor based on the second convolutional neural network model, and the decoder;
wherein the training phase comprises:
acquiring training data, wherein the training data comprises training fluorescence development images of detected samples and a true value of the number of CTC cells;
performing image preprocessing on the training fluorescence development image to obtain a fluorescence development image after training preprocessing;
the fluorescence development image after training pretreatment passes through the shallow feature extractor based on the first convolutional neural network model to obtain a training fluorescence shallow feature map;
The training fluorescence shallow feature map passes through the spatial attention module to obtain a training space enhanced fluorescence shallow feature map;
passing the training space enhanced fluorescence shallow feature map through the deep feature extractor based on the second convolutional neural network model to obtain a training fluorescence deep feature map;
fusing the training space enhanced fluorescence shallow feature map and the training fluorescence deep feature map to obtain a training fluorescence feature map;
optimizing the characteristic manifold curved surface of the training fluorescence characteristic map to obtain an optimized fluorescence characteristic map;
passing the optimized fluorescence signature through the decoder to obtain a training decoded image;
performing image semantic segmentation on the decoded image to obtain a segmentation result, and determining the number of CTC cells based on the segmentation result; and
calculating the mean square error between the number of CTC cells and the true value of the number of CTC cells as a loss function value, and training the shallow feature extractor based on the first convolutional neural network model, the spatial attention module, the deep feature extractor based on the second convolutional neural network model and the decoder through the propagation direction of gradient descent.
10. The fluorescence microscopy imaging method for detection of circulating tumor cells of claim 9, wherein feature manifold surface optimization of the training fluorescence signature to obtain an optimized fluorescence signature comprises: optimizing the characteristic manifold curved surface of the training fluorescence characteristic map by using the following optimization formula to obtain the optimized fluorescence characteristic map;
wherein, the optimization formula is:
wherein the method comprises the steps ofIs the +.f. of the training fluorescence profile>Characteristic value of individual position->And->Is the mean and standard deviation of the feature value sets of each position in the training fluorescence feature map, +.>Represents the maximum function, and->Is the +.f. of the optimized fluorescence profile>Characteristic values of the location.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202310581412.8A CN116363123B (en) | 2023-05-23 | 2023-05-23 | Fluorescence microscopic imaging system and method for detecting circulating tumor cells |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202310581412.8A CN116363123B (en) | 2023-05-23 | 2023-05-23 | Fluorescence microscopic imaging system and method for detecting circulating tumor cells |
Publications (2)
Publication Number | Publication Date |
---|---|
CN116363123A true CN116363123A (en) | 2023-06-30 |
CN116363123B CN116363123B (en) | 2023-12-22 |
Family
ID=86922509
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202310581412.8A Active CN116363123B (en) | 2023-05-23 | 2023-05-23 | Fluorescence microscopic imaging system and method for detecting circulating tumor cells |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN116363123B (en) |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN116586924A (en) * | 2023-07-17 | 2023-08-15 | 浙江一益医疗器械有限公司 | Stainless steel needle tube with needle tip five-inclined-plane structure and preparation process thereof |
CN116612472A (en) * | 2023-07-21 | 2023-08-18 | 北京航空航天大学杭州创新研究院 | Single-molecule immune array analyzer based on image and method thereof |
CN116630313A (en) * | 2023-07-21 | 2023-08-22 | 北京航空航天大学杭州创新研究院 | Fluorescence imaging detection system and method thereof |
CN117392134A (en) * | 2023-12-12 | 2024-01-12 | 苏州矩度电子科技有限公司 | On-line visual detection system for high-speed dispensing |
CN117475241A (en) * | 2023-12-27 | 2024-01-30 | 山西省水利建筑工程局集团有限公司 | Geological mutation detection system and method for tunnel excavation of cantilever type heading machine |
CN118365890A (en) * | 2024-06-20 | 2024-07-19 | 西安交通大学医学院第一附属医院 | Intelligent segmentation system and method for digital pathological image tissue of non-small cell lung cancer |
Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107748256A (en) * | 2017-10-11 | 2018-03-02 | 上海医盈网络科技有限公司 | A kind of liquid biopsy detection method of circulating tumor cell |
CN109886273A (en) * | 2019-02-26 | 2019-06-14 | 四川大学华西医院 | A kind of CMR classification of image segmentation system |
CN113688826A (en) * | 2021-07-05 | 2021-11-23 | 北京工业大学 | Pollen image detection method and system based on feature fusion |
US11222217B1 (en) * | 2020-08-14 | 2022-01-11 | Tsinghua University | Detection method using fusion network based on attention mechanism, and terminal device |
WO2022257408A1 (en) * | 2021-06-10 | 2022-12-15 | 南京邮电大学 | Medical image segmentation method based on u-shaped network |
CN115619320A (en) * | 2022-11-09 | 2023-01-17 | 西安巨子生物基因技术股份有限公司 | Product external packing informatization error-preventing system |
US20230036359A1 (en) * | 2020-01-07 | 2023-02-02 | Raycan Technology Co., Ltd. (Suzhou) | Image reconstruction method, device,equipment, system, and computer-readable storage medium |
CN115868923A (en) * | 2022-04-21 | 2023-03-31 | 华中科技大学 | Fluorescence molecule tomography method and system based on expanded cyclic neural network |
WO2023070447A1 (en) * | 2021-10-28 | 2023-05-04 | 京东方科技集团股份有限公司 | Model training method, image processing method, computing processing device, and non-transitory computer readable medium |
-
2023
- 2023-05-23 CN CN202310581412.8A patent/CN116363123B/en active Active
Patent Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107748256A (en) * | 2017-10-11 | 2018-03-02 | 上海医盈网络科技有限公司 | A kind of liquid biopsy detection method of circulating tumor cell |
CN109886273A (en) * | 2019-02-26 | 2019-06-14 | 四川大学华西医院 | A kind of CMR classification of image segmentation system |
US20230036359A1 (en) * | 2020-01-07 | 2023-02-02 | Raycan Technology Co., Ltd. (Suzhou) | Image reconstruction method, device,equipment, system, and computer-readable storage medium |
US11222217B1 (en) * | 2020-08-14 | 2022-01-11 | Tsinghua University | Detection method using fusion network based on attention mechanism, and terminal device |
WO2022257408A1 (en) * | 2021-06-10 | 2022-12-15 | 南京邮电大学 | Medical image segmentation method based on u-shaped network |
CN113688826A (en) * | 2021-07-05 | 2021-11-23 | 北京工业大学 | Pollen image detection method and system based on feature fusion |
WO2023070447A1 (en) * | 2021-10-28 | 2023-05-04 | 京东方科技集团股份有限公司 | Model training method, image processing method, computing processing device, and non-transitory computer readable medium |
CN115868923A (en) * | 2022-04-21 | 2023-03-31 | 华中科技大学 | Fluorescence molecule tomography method and system based on expanded cyclic neural network |
CN115619320A (en) * | 2022-11-09 | 2023-01-17 | 西安巨子生物基因技术股份有限公司 | Product external packing informatization error-preventing system |
Non-Patent Citations (3)
Title |
---|
DAWEI YANG, ET AL.: "Image semantic segmentation with hierarchical feature fusion based on deep neural network", 《CONNECTION SCIENCE》, pages 158 * |
周鹏 等: "基于改进 DeepLab-v3 +的火星地形分割算法", 《空间控制技术与应用》 * |
蒋欣欣: "结合实例分割的多目标跟踪研究及系统实现", 《中国优秀硕士学位论文全文数据库 信息科技辑》, no. 1, pages 4 * |
Cited By (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN116586924A (en) * | 2023-07-17 | 2023-08-15 | 浙江一益医疗器械有限公司 | Stainless steel needle tube with needle tip five-inclined-plane structure and preparation process thereof |
CN116586924B (en) * | 2023-07-17 | 2024-02-27 | 浙江一益医疗器械有限公司 | Stainless steel needle tube with needle tip five-inclined-plane structure and preparation process thereof |
CN116612472A (en) * | 2023-07-21 | 2023-08-18 | 北京航空航天大学杭州创新研究院 | Single-molecule immune array analyzer based on image and method thereof |
CN116630313A (en) * | 2023-07-21 | 2023-08-22 | 北京航空航天大学杭州创新研究院 | Fluorescence imaging detection system and method thereof |
CN116612472B (en) * | 2023-07-21 | 2023-09-19 | 北京航空航天大学杭州创新研究院 | Single-molecule immune array analyzer based on image and method thereof |
CN116630313B (en) * | 2023-07-21 | 2023-09-26 | 北京航空航天大学杭州创新研究院 | Fluorescence imaging detection system and method thereof |
CN117392134A (en) * | 2023-12-12 | 2024-01-12 | 苏州矩度电子科技有限公司 | On-line visual detection system for high-speed dispensing |
CN117392134B (en) * | 2023-12-12 | 2024-02-27 | 苏州矩度电子科技有限公司 | On-line visual detection system for high-speed dispensing |
CN117475241A (en) * | 2023-12-27 | 2024-01-30 | 山西省水利建筑工程局集团有限公司 | Geological mutation detection system and method for tunnel excavation of cantilever type heading machine |
CN118365890A (en) * | 2024-06-20 | 2024-07-19 | 西安交通大学医学院第一附属医院 | Intelligent segmentation system and method for digital pathological image tissue of non-small cell lung cancer |
Also Published As
Publication number | Publication date |
---|---|
CN116363123B (en) | 2023-12-22 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN116363123B (en) | Fluorescence microscopic imaging system and method for detecting circulating tumor cells | |
CN116189179B (en) | Circulating tumor cell scanning analysis equipment | |
CN111488921B (en) | Intelligent analysis system and method for panoramic digital pathological image | |
JP7273215B2 (en) | Automated assay evaluation and normalization for image processing | |
CN113344849B (en) | Microemulsion head detection system based on YOLOv5 | |
US10621412B2 (en) | Dot detection, color classification of dots and counting of color classified dots | |
CN112926652B (en) | Fish fine granularity image recognition method based on deep learning | |
CN113393443B (en) | HE pathological image cell nucleus segmentation method and system | |
CN117015796A (en) | Method for processing tissue images and system for processing tissue images | |
CN110853011A (en) | Method for constructing convolutional neural network model for pulmonary nodule detection | |
CN115410050A (en) | Tumor cell detection equipment based on machine vision and method thereof | |
CN112990015B (en) | Automatic identification method and device for lesion cells and electronic equipment | |
CN112703531A (en) | Generating annotation data for tissue images | |
CN116287138B (en) | FISH-based cell detection system and method thereof | |
CN111680575B (en) | Human epithelial cell staining classification device, equipment and storage medium | |
CN114581698A (en) | Target classification method based on space cross attention mechanism feature fusion | |
CN116843974A (en) | Breast cancer pathological image classification method based on residual neural network | |
Abbasi-Sureshjani et al. | Molecular subtype prediction for breast cancer using H&E specialized backbone | |
CN117746077A (en) | Chip defect detection method, device, equipment and storage medium | |
CN113837255B (en) | Method, apparatus and medium for predicting cell-based antibody karyotype class | |
CN113536896B (en) | Insulator defect detection method and device based on improved Faster RCNN and storage medium | |
CN112991281B (en) | Visual detection method, system, electronic equipment and medium | |
CN116843956A (en) | Cervical pathology image abnormal cell identification method, system and storage medium | |
CN115131628A (en) | Mammary gland image classification method and equipment based on typing auxiliary information | |
KR20230063147A (en) | Efficient Lightweight CNN and Ensemble Machine Learning Classification of Prostate Tissue Using Multilevel Feature Analysis Method and System |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |