WO2022247296A1 - Mark point-based image registration method - Google Patents

Mark point-based image registration method Download PDF

Info

Publication number
WO2022247296A1
WO2022247296A1 PCT/CN2022/070425 CN2022070425W WO2022247296A1 WO 2022247296 A1 WO2022247296 A1 WO 2022247296A1 CN 2022070425 W CN2022070425 W CN 2022070425W WO 2022247296 A1 WO2022247296 A1 WO 2022247296A1
Authority
WO
WIPO (PCT)
Prior art keywords
image
matrix
matching
matching point
rigid registration
Prior art date
Application number
PCT/CN2022/070425
Other languages
French (fr)
Chinese (zh)
Inventor
魏军
沈烁
卢旭玲
朱德明
田孟秋
Original Assignee
广州柏视医疗科技有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 广州柏视医疗科技有限公司 filed Critical 广州柏视医疗科技有限公司
Publication of WO2022247296A1 publication Critical patent/WO2022247296A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/30Determination of transform parameters for the alignment of images, i.e. image registration
    • G06T7/33Determination of transform parameters for the alignment of images, i.e. image registration using feature-based methods
    • G06T7/337Determination of transform parameters for the alignment of images, i.e. image registration using feature-based methods involving reference images or patches
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • G06T7/0012Biomedical image inspection
    • G06T7/0014Biomedical image inspection using an image reference approach
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10072Tomographic images
    • G06T2207/10084Hybrid tomography; Concurrent acquisition with multiple different tomographic modalities
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]

Definitions

  • the invention relates to the field of image processing, deep learning, and medical treatment, in particular to an image registration method based on marker points.
  • Image registration has many practical applications in medical image processing and analysis. With the advancement of medical imaging equipment, for the same patient, images of multiple different modalities containing accurate anatomical information can be collected, such as CT, CBCT, MRI, PET, etc. However, making a diagnosis by observing different images requires spatial imagination and the doctor's subjective experience. With the correct image registration method, a variety of information can be accurately fused into the same image, making it easier and more accurate for doctors to observe lesions and structures from various angles. At the same time, through the registration of dynamic images collected at different times, the changes in lesions and organs can be quantitatively analyzed, making medical diagnosis, surgical planning, and radiotherapy planning more accurate and reliable.
  • the traditional image registration method is based on the optimization of the similarity objective function to solve the problem, which is easy to converge to the local minimum, especially the registration effect on different modal images is poor, and the iterative solution process takes a long time.
  • the image registration method based on markers can solve the above problems, but the acquisition of the gold standard of markers requires a lot of time for doctors and experts, and the cost is high.
  • AI algorithms have been used to build mathematical models that outperform human medical experts. Therefore, there is reason to believe that using AI algorithms to improve traditional image registration methods can effectively improve the effect of image registration.
  • the object of the present invention is to provide an image registration method based on marker points, which can improve the traditional image registration method by using AI algorithm and effectively improve the effect of image registration.
  • the present invention provides a marker-based image registration method, which mainly includes the following steps: input two medical images of any modality (CT, CBCT, MRI, PET, etc.), one as a fixed image (reference image), and the other as a moving image (image to be registered); a pre-trained neural network is used to extract the pyramid features of the two input images.
  • CT computed tomography
  • CBCT computed tomography
  • MRI magnetic resonance imaging
  • PET etc.
  • a pre-trained neural network is used to extract the pyramid features of the two input images.
  • the training process of the network includes many different tasks and involves the above-mentioned many different input mode; use the above-mentioned neural network to extract pyramid features, and obtain multiple matching point pairs representing certain semantics between two images through searching, screening, matching and other processes; by minimizing the sum of the distances between all matching point pairs , to fit the transformation matrix and displacement vector of rigid registration, so as to obtain the warped image of the medical image after rigid registration. And on the basis of rigid registration, the three-dimensional matrix of the displacement field of non-rigid registration is obtained by interpolation method based on radial basis, so as to obtain the warped image of the medical image after non-rigid registration.
  • a pre-trained neural network is used to extract the pyramid features of two input images, including: the structure of the neural network is divided into a backbone network and a plurality of subsequent branch networks.
  • the backbone network is shared among different tasks, and each branch network corresponds to a task.
  • the backbone network is used to extract image features.
  • the training process of the neural network involves a variety of different tasks and involves a variety of different input modalities, including but not limited to: CT-based NPC primary tumor (GTV) segmentation, MRI-based NPC primary tumor segmentation , CT-based cervical cancer primary tumor segmentation, PET-based lung primary tumor segmentation, CT-based organ-at-risk (OAR) segmentation, MRI-based organ-at-risk segmentation, CBCT-based organ-at-risk segmentation, CT-based lung Nodule object detection, etc. And first use one of the tasks to train the neural network, and then train each of the other input modal tasks at the same time, and then train each of the remaining tasks separately.
  • the backbone network parameters are fixed during training, and finally all tasks are trained at the same time to fine-tune all parameter.
  • the image registration method based on markers further includes: using the above-mentioned neural network to extract pyramid features, and obtaining multiple matching points representing certain semantics between two images through processes such as searching, screening, and matching Yes, including the following steps: input If (fixed image) and I m (moving image) into the above-mentioned pre-trained neural network, and extract the pyramidal feature map (feature map) of the two input images and Among them, l ⁇ 1,2,3,4,5 ⁇ represents the l-level feature, and the larger the number, the deeper the layer, that is, the smaller the feature size is, the more high-level semantics it contains.
  • N1 is the number of the lth level search range
  • S1 is a set of multiple search range pairs of the lth level.
  • ⁇ (p l ) is the point set of the neighborhood within a specific range of point p l .
  • the image registration method based on marker points further includes: by minimizing the sum of the distances between all matching point pairs, fitting the transformation matrix and displacement vector of rigid registration, thereby obtaining rigid registration
  • the warped image of the final medical image includes the following steps: After obtaining all matching point pairs, the optimal solution of the transformation matrix and displacement vector of rigid registration is obtained by minimizing the following formula:
  • N is the number of matching point pairs
  • p n is the nth matching point of the fixed image
  • q n is the pixel point in the corresponding moving image.
  • P is a matrix composed of all matching points of the fixed image, with a size of [N,4], that is, a matrix composed of N four-dimensional row vectors. The first three dimensions of the four dimensions are the physical coordinates of the pixels, and the fourth dimension is a fixed value of 1.
  • Q is a matrix composed of all matching points of the moving image, with a size of [N, 4].
  • the size of matrix R is [4,4], R[0:3,0:3] refers to the matrix of size [3,3] composed of the first 3 rows and first 3 columns of matrix R, R[0:3 ,3] refers to the three-dimensional column vector of the first 3 rows and the 3rd column of the matrix R.
  • a and b are the optimal solution of transformation matrix and displacement vector respectively.
  • the warped image of the medical image after rigid registration is obtained through A and b.
  • the marker-based image registration method further includes: on the basis of rigid registration, obtain the three-dimensional displacement field matrix of non-rigid registration by interpolation method based on radial basis, so as to obtain the non-rigid registration
  • the warped image of the medical image after registration includes the following steps: the size of the three-dimensional matrix of the displacement field is the same as that of the fixed image. After obtaining N matching point pairs, use the following interpolation method to obtain the value of the remaining pixels of the displacement field matrix:
  • A (a 1 ,a 2 ,a 3 )
  • V (v 1 ,v 2 ,...,v N ,0,0,0,0)
  • (w 1 ,w 2 ,...,w N ,b,a 1 ,a 2 ,a 3 )
  • P is a matrix composed of all matching points of the fixed image, with a size of [N, 4], that is, a matrix composed of n four-dimensional row vectors.
  • the last three dimensions of the four dimensions are the physical coordinates of the pixels, and the first dimension is a fixed value of 1.
  • q n is the corresponding matching point of p n in the moving image.
  • the displacement values of the pixel points in the displacement field except the matching points can be obtained by f() fitting. Since the displacement value is a three-dimensional vector, that is, the x, y, and z directions, the above interpolation process needs to be repeated three times, that is, each direction is performed once. And finally obtain the three-dimensional matrix of the displacement field, so as to obtain the warped image of the medical image after non-rigid registration.
  • the marker-based image registration method of the present invention has the following beneficial effects: (1) The present invention can register images of any two modalities. (2) The present invention adopts a pre-trained neural network to extract image features. The training process of this network includes multiple different tasks and involves multiple different input modes, which can effectively improve the effectiveness and versatility of features. (3) The present invention utilizes a pre-trained neural network to extract image features, and obtains multiple matching point pairs representing certain semantics between two images through processes such as search, screening, and matching, which can effectively solve the problem of lack of gold standard for marking points question. (4) The present invention realizes rigid registration by minimizing the sum of the distances between all matching point pairs and solving the transformation matrix and displacement vector. (5) The present invention solves the three-dimensional matrix of the displacement field through the interpolation method based on the radial basis to realize non-rigid registration.
  • Fig. 1 is a schematic flow chart of an image registration method based on markers according to an embodiment of the present invention.
  • a kind of image registration method based on marker point comprises the following steps:
  • Pyramid features of two input images are extracted using a pre-trained neural network, which is trained on a variety of tasks and involves the various input modalities mentioned above;
  • the transformation matrix and displacement vector for rigid registration are fitted to obtain the warped image of the medical image after rigid registration.
  • the three-dimensional matrix of the displacement field of non-rigid registration is obtained by interpolation method based on radial basis, so as to obtain the warped image of the medical image after non-rigid registration.
  • a specific implementation workflow of a marker-based image registration method of the present invention includes:
  • step S1 constructing a pre-trained neural network to extract the pyramid features of two input images
  • Step S1 specifically includes the following steps:
  • the structure of the neural network is divided into a backbone network and a plurality of subsequent branch networks.
  • the backbone network is shared among different tasks, and each branch network corresponds to a task.
  • the backbone network is used to extract image features.
  • the training process of the neural network involves a variety of different tasks and involves a variety of different input modalities, including but not limited to: CT-based nasopharyngeal carcinoma primary tumor (GTV) segmentation, MRI-based nasopharyngeal carcinoma primary Tumor segmentation, CT-based cervical cancer primary tumor segmentation, PET-based lung primary tumor segmentation, CT-based organ-at-risk (OAR) segmentation, MRI-based organ-at-risk segmentation, CBCT-based organ-at-risk segmentation, CT-based Pulmonary nodule target detection, etc.
  • GTV nasopharyngeal carcinoma primary tumor
  • OAR organ-at-risk
  • MRI-based organ-at-risk segmentation CBCT-based organ-at-risk segmentation
  • CT-based Pulmonary nodule target detection etc.
  • the image registration method based on markers also includes:
  • Step S2 specifically includes the following steps:
  • N1 is the number of the lth level search range
  • S1 is a set of multiple search range pairs of the lth level.
  • ⁇ (p l ) is the point set of the neighborhood within a specific range of point p l .
  • is a custom threshold, which is 0.05 here.
  • step S28 After obtaining the upper-level search range set, for each search range, jump to step S23 and repeat steps S23-S28 to obtain the final output result That is, the set of matching point pairs of the two images.
  • the image registration method based on markers also includes:
  • Step S3 specifically includes the following steps:
  • N is the number of matching point pairs
  • p n is the nth matching point of the fixed image
  • q n is the pixel point in the corresponding moving image.
  • P is a matrix composed of all matching points of the fixed image, with a size of [N,4], that is, a matrix composed of N four-dimensional row vectors. The first three dimensions of the four dimensions are the physical coordinates of the pixels, and the fourth dimension is a fixed value of 1.
  • Q is a matrix composed of all matching points of the moving image, with a size of [N, 4].
  • the size of matrix R is [4,4], R[0:3,0:3] refers to the matrix of size [3,3] composed of the first 3 rows and first 3 columns of matrix R, R[0:3 ,3] refers to the three-dimensional column vector of the first 3 rows and the 3rd column of the matrix R.
  • a and b are the optimal solution of transformation matrix and displacement vector respectively.
  • the image registration method based on markers also includes:
  • the three-dimensional matrix of the displacement field of non-rigid registration is obtained by interpolation method based on radial basis, so as to obtain the warped image of the medical image after non-rigid registration;
  • Step S4 specifically includes the following steps:
  • the size of the displacement field three-dimensional matrix is the same as that of the fixed image. After obtaining N matching point pairs, use the following interpolation method to obtain the value of the remaining pixels of the displacement field matrix:
  • A (a 1 ,a 2 ,a 3 )
  • V (v 1 ,v 2 ,...,v N ,0,0,0,0)
  • (w 1 ,w 2 ,...,w N ,b,a 1 ,a 2 ,a 3 )
  • P is a matrix composed of all matching points of the fixed image, with a size of [N, 4], that is, a matrix composed of n four-dimensional row vectors.
  • the last three dimensions of the four dimensions are the physical coordinates of the pixels, and the first dimension is a fixed value of 1.
  • q n is the corresponding matching point of p n in the moving image.
  • the displacement values of the pixel points in the displacement field except the matching points can be obtained by f() fitting.
  • step S41 is performed once for each direction.
  • the marker-based image registration method of the present invention has the following advantages: (1) The present invention can register images of any two modalities. (2) The present invention adopts a pre-trained neural network to extract image features, and the training process of this network includes multiple different tasks and involves multiple different input modes, which can effectively improve the effectiveness and versatility of features. (3) The present invention utilizes a pre-trained neural network to extract image features, and obtains multiple matching point pairs representing certain semantics between two images through processes such as search, screening, and matching, which can effectively solve the problem of lack of gold standard for marking points question. (4) The present invention realizes rigid registration by minimizing the sum of the distances between all matching point pairs and solving the transformation matrix and displacement vector. (5) The present invention solves the three-dimensional matrix of the displacement field through the interpolation method based on the radial basis to realize non-rigid registration.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Data Mining & Analysis (AREA)
  • Radiology & Medical Imaging (AREA)
  • Quality & Reliability (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Medical Informatics (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
  • Evolutionary Computation (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Image Analysis (AREA)
  • Image Processing (AREA)

Abstract

A mark point-based image registration method, comprising: inputting two medical images in any mode; extracting pyramid features of the two input images by using a pre-trained neural network, wherein the training process of the network comprises a plurality of different tasks and relates to a plurality of different input modes; extracting the pyramid features by using the neural network, and obtaining a plurality of matching point pairs representing certain semantics between the two images by means of the processes of searching, screening, matching and the like; fitting a transformation matrix and a displacement vector of rigid registration by minimizing the sum of point distances between all the matching point pairs, to obtain a medical image Warped image after the rigid registration; and on the basis of the rigid registration, obtaining a displacement field three-dimensional matrix of non-rigid registration by means of an interpolation method based on a radial basis, to obtain a medical image Warped image after the non-rigid registration. Therefore, the method can effectively improve the image registration effect.

Description

基于标记点的图像配准方法Image registration method based on marker points 技术领域technical field
本发明涉及图像处理领域、深度学习领域、医疗领域,尤其是一种基于标记点的图像配准方法。The invention relates to the field of image processing, deep learning, and medical treatment, in particular to an image registration method based on marker points.
背景技术Background technique
图像配准在医学图像处理与分析中有众多具有实用价值的应用。随着医学成像设备的进步,对于同一患者,可以采集含有准确解剖信息的多种不同模态的图像,如CT、CBCT、MRI、PET等。然而,通过观察不同图像进行诊断需要凭着空间想象和医生的主观经验。采用正确的图像配准方法则可以将多种多样的信息准确地融合到同一图像中,使医生更方便更精确地从各个角度观察病灶和结构。同时,通过对不同时刻采集的动态图像的配准,可以定量分析病灶和器官的变化情况,使得医疗诊断、制定手术计划、放射治疗计划更准确可靠。Image registration has many practical applications in medical image processing and analysis. With the advancement of medical imaging equipment, for the same patient, images of multiple different modalities containing accurate anatomical information can be collected, such as CT, CBCT, MRI, PET, etc. However, making a diagnosis by observing different images requires spatial imagination and the doctor's subjective experience. With the correct image registration method, a variety of information can be accurately fused into the same image, making it easier and more accurate for doctors to observe lesions and structures from various angles. At the same time, through the registration of dynamic images collected at different times, the changes in lesions and organs can be quantitatively analyzed, making medical diagnosis, surgical planning, and radiotherapy planning more accurate and reliable.
传统的图像配准方法基于相似度目标函数的优化求解问题,容易收敛至局部极小值,尤其对不同模态图像的配准效果较差,且迭代求解的过程耗时较长。而基于标记点的图像配准方法能解决上述问题,但标记点金标准的获得需要耗费医生、专家的大量时间,成本较高。近年来,人们对探索利用人工智进行诊断产生了浓厚的兴趣,并在某些领域利用AI算法建立了表现优于人类医学专家的数学模型。因此有理由相信,利用AI算法对传统图像配准方法进行改进能有效提高图像配准的效果。The traditional image registration method is based on the optimization of the similarity objective function to solve the problem, which is easy to converge to the local minimum, especially the registration effect on different modal images is poor, and the iterative solution process takes a long time. The image registration method based on markers can solve the above problems, but the acquisition of the gold standard of markers requires a lot of time for doctors and experts, and the cost is high. In recent years, there has been a lot of interest in exploring the use of artificial intelligence for diagnosis, and in some areas AI algorithms have been used to build mathematical models that outperform human medical experts. Therefore, there is reason to believe that using AI algorithms to improve traditional image registration methods can effectively improve the effect of image registration.
公开于该背景技术部分的信息仅仅旨在增加对本发明的总体背景的理解,而不应当被视为承认或以任何形式暗示该信息构成已为本领域一般技术人员所公知的现有技术。The information disclosed in this Background section is only for enhancing the understanding of the general background of the present invention and should not be taken as an acknowledgment or any form of suggestion that the information constitutes the prior art that is already known to those skilled in the art.
发明内容Contents of the invention
本发明的目的在于提供一种基于标记点的图像配准方法,其能够利用AI算法对传统图像配准方法进行改进能有效提高图像配准的效果。The object of the present invention is to provide an image registration method based on marker points, which can improve the traditional image registration method by using AI algorithm and effectively improve the effect of image registration.
为实现上述目的,本发明提供了一种基于标记点的图像配准方法,其主要包括下列步骤:输入两个任意模态(CT、CBCT、MRI、PET等)的医学图像,一个作为fixed image(参考图像),另一个作为moving image(待配准图像);采用一个预训练的神经网络提取两个输入图像的金字塔特征,该网络的训练过程包含多种不同的任务并涉及上述多种不同的输入模态;利用上述神经网络提取金字塔特征,通过搜索、筛选、匹配等过程得到两个图像间的代表某种语义的多个匹配点对;通过最小化所有匹配点对间点距离的总和,拟合出刚性配准的变换矩阵和位移向量,从而得到刚性配准后的医学图像warped image。以及在刚性配准的基础上,通过基于径向基的插值法得到非刚性配准的位移场三维矩阵,从而得到非刚性配准后的医学图像warped image。In order to achieve the above object, the present invention provides a marker-based image registration method, which mainly includes the following steps: input two medical images of any modality (CT, CBCT, MRI, PET, etc.), one as a fixed image (reference image), and the other as a moving image (image to be registered); a pre-trained neural network is used to extract the pyramid features of the two input images. The training process of the network includes many different tasks and involves the above-mentioned many different input mode; use the above-mentioned neural network to extract pyramid features, and obtain multiple matching point pairs representing certain semantics between two images through searching, screening, matching and other processes; by minimizing the sum of the distances between all matching point pairs , to fit the transformation matrix and displacement vector of rigid registration, so as to obtain the warped image of the medical image after rigid registration. And on the basis of rigid registration, the three-dimensional matrix of the displacement field of non-rigid registration is obtained by interpolation method based on radial basis, so as to obtain the warped image of the medical image after non-rigid registration.
在一优选的实施方式中,采用一个预训练的神经网络提取两个输入图像的金字塔特征,包括:神经网络的结构分为骨干网络和后续多个分支网络。骨干网络在不同任务间共享,每个分支网络对应一个任务。最后用于提取图像特征的是骨干网络。神经网络的训练过程包含多种不同的任务并涉及多种不同的输入模态,包括但不限于:基于CT的鼻咽癌原发肿瘤(GTV)分割、基于MRI的鼻咽癌原发肿瘤分割、基于CT的子宫颈癌原发肿瘤分割、基于PET的肺部原发肿瘤分割、基于CT的危及器官(OAR)分割、基于MRI的危及器官分割、基于CBCT的危及器官分割、基于CT的肺结节目标检测等。以及先用其中一个任务训练神经网络,再加上其它输入模态的任务各一个同时训练,再分别对剩余的每个任务单独进行训练,训练时骨干网络参数固定,最后所有任务同时训练微调所有参数。In a preferred embodiment, a pre-trained neural network is used to extract the pyramid features of two input images, including: the structure of the neural network is divided into a backbone network and a plurality of subsequent branch networks. The backbone network is shared among different tasks, and each branch network corresponds to a task. Finally, the backbone network is used to extract image features. The training process of the neural network involves a variety of different tasks and involves a variety of different input modalities, including but not limited to: CT-based NPC primary tumor (GTV) segmentation, MRI-based NPC primary tumor segmentation , CT-based cervical cancer primary tumor segmentation, PET-based lung primary tumor segmentation, CT-based organ-at-risk (OAR) segmentation, MRI-based organ-at-risk segmentation, CBCT-based organ-at-risk segmentation, CT-based lung Nodule object detection, etc. And first use one of the tasks to train the neural network, and then train each of the other input modal tasks at the same time, and then train each of the remaining tasks separately. The backbone network parameters are fixed during training, and finally all tasks are trained at the same time to fine-tune all parameter.
在一优选的实施方式中,基于标记点的图像配准方法还包括:利用上述神经网络提取金字塔特征,通过搜索、筛选、匹配等过程得到两个图像间的 代表某种语义的多个匹配点对,包括以下步骤:把I f(fixed image)与I m(moving image)输入上述预训练的神经网络,提取两个输入图像的金字塔特征图(feature map)
Figure PCTCN2022070425-appb-000001
Figure PCTCN2022070425-appb-000002
其中l∈{1,2,3,4,5}表示第l级特征,数字越大表示层数越深,即特征尺寸越小但蕴含更多高层语义。匹配点的搜索需要在特定的搜索范围内产生,从l=5开始设置搜索范围:
In a preferred embodiment, the image registration method based on markers further includes: using the above-mentioned neural network to extract pyramid features, and obtaining multiple matching points representing certain semantics between two images through processes such as searching, screening, and matching Yes, including the following steps: input If (fixed image) and I m (moving image) into the above-mentioned pre-trained neural network, and extract the pyramidal feature map (feature map) of the two input images
Figure PCTCN2022070425-appb-000001
and
Figure PCTCN2022070425-appb-000002
Among them, l∈{1,2,3,4,5} represents the l-level feature, and the larger the number, the deeper the layer, that is, the smaller the feature size is, the more high-level semantics it contains. The search for matching points needs to be generated within a specific search range, and the search range is set starting from l=5:
S 5={(P 5,Q 5)} S 5 ={(P 5 ,Q 5 )}
其中:in:
Figure PCTCN2022070425-appb-000003
Figure PCTCN2022070425-appb-000003
Figure PCTCN2022070425-appb-000004
为I f的第l级的第n个搜索范围,对应地
Figure PCTCN2022070425-appb-000005
为I m的第l级的第n个搜索范围,N l为第l级的搜索范围的个数,S l为第l级的多个搜索范围对的集合。当l=5时,搜索范围是
Figure PCTCN2022070425-appb-000006
Figure PCTCN2022070425-appb-000007
的整个范围,即N 5=1。
Figure PCTCN2022070425-appb-000004
is the nth search range of the lth level of If , correspondingly
Figure PCTCN2022070425-appb-000005
is the nth search range of the lth level of Im , N1 is the number of the lth level search range, and S1 is a set of multiple search range pairs of the lth level. When l=5, the search range is
Figure PCTCN2022070425-appb-000006
and
Figure PCTCN2022070425-appb-000007
The entire range of , that is, N 5 =1.
通过下式对搜索范围
Figure PCTCN2022070425-appb-000008
Figure PCTCN2022070425-appb-000009
的特征图
Figure PCTCN2022070425-appb-000010
Figure PCTCN2022070425-appb-000011
进行变换,得到
Figure PCTCN2022070425-appb-000012
Figure PCTCN2022070425-appb-000013
Search range by the following formula
Figure PCTCN2022070425-appb-000008
and
Figure PCTCN2022070425-appb-000009
feature map of
Figure PCTCN2022070425-appb-000010
and
Figure PCTCN2022070425-appb-000011
transform to get
Figure PCTCN2022070425-appb-000012
and
Figure PCTCN2022070425-appb-000013
Figure PCTCN2022070425-appb-000014
Figure PCTCN2022070425-appb-000014
Figure PCTCN2022070425-appb-000015
Figure PCTCN2022070425-appb-000015
Figure PCTCN2022070425-appb-000016
Figure PCTCN2022070425-appb-000016
Figure PCTCN2022070425-appb-000017
Figure PCTCN2022070425-appb-000017
其中,
Figure PCTCN2022070425-appb-000018
表示在
Figure PCTCN2022070425-appb-000019
范围内的局部特征图,
Figure PCTCN2022070425-appb-000020
为变换后的特征图,
Figure PCTCN2022070425-appb-000021
Figure PCTCN2022070425-appb-000022
的均值,
Figure PCTCN2022070425-appb-000023
Figure PCTCN2022070425-appb-000024
的标准差,同
Figure PCTCN2022070425-appb-000025
Figure PCTCN2022070425-appb-000026
in,
Figure PCTCN2022070425-appb-000018
expressed in
Figure PCTCN2022070425-appb-000019
Local feature maps in the range,
Figure PCTCN2022070425-appb-000020
is the transformed feature map,
Figure PCTCN2022070425-appb-000021
for
Figure PCTCN2022070425-appb-000022
the mean value of
Figure PCTCN2022070425-appb-000023
for
Figure PCTCN2022070425-appb-000024
standard deviation of
Figure PCTCN2022070425-appb-000025
Figure PCTCN2022070425-appb-000026
在搜索范围
Figure PCTCN2022070425-appb-000027
Figure PCTCN2022070425-appb-000028
内搜索匹配点对,当满足以下条件时,两个点p l与q l为 匹配点对:
in the search range
Figure PCTCN2022070425-appb-000027
and
Figure PCTCN2022070425-appb-000028
Search for matching point pairs within, when the following conditions are met, the two points p l and q l are matching point pairs:
Figure PCTCN2022070425-appb-000029
Figure PCTCN2022070425-appb-000029
Figure PCTCN2022070425-appb-000030
Figure PCTCN2022070425-appb-000030
Figure PCTCN2022070425-appb-000031
Figure PCTCN2022070425-appb-000031
即若以
Figure PCTCN2022070425-appb-000032
内的点p l
Figure PCTCN2022070425-appb-000033
范围内搜索时,相似度最高的点是q l,反之也成立,则p l与q l为匹配点对。相似度的计算公式为:
That is to say, if
Figure PCTCN2022070425-appb-000032
The point p l inside
Figure PCTCN2022070425-appb-000033
When searching within the range, the point with the highest similarity is q l , and vice versa, then p l and q l are matching point pairs. The formula for calculating the similarity is:
Figure PCTCN2022070425-appb-000034
Figure PCTCN2022070425-appb-000034
其中,ε(p l)为点p l的特定范围内邻域的点集。 Among them, ε(p l ) is the point set of the neighborhood within a specific range of point p l .
第l级的所有N l个搜索范围分别执行上述搜索匹配点的步骤,便得到所有匹配点对的集合Λ lAll the N1 search ranges of the lth level perform the above-mentioned steps of searching for matching points respectively, and then obtain the set Λl of all matching point pairs:
Figure PCTCN2022070425-appb-000035
Figure PCTCN2022070425-appb-000035
对于上述步骤搜索得到的匹配点对,还必须通过以下的筛选条件,即该点在特征图(feature map)的值必须足够大:For the matching point pairs searched in the above steps, the following filter conditions must also be passed, that is, the value of the point in the feature map (feature map) must be large enough:
Figure PCTCN2022070425-appb-000036
Figure PCTCN2022070425-appb-000036
Figure PCTCN2022070425-appb-000037
Figure PCTCN2022070425-appb-000037
其中,
Figure PCTCN2022070425-appb-000038
为最终得到的匹配点对集合,γ为自定义阈值。
in,
Figure PCTCN2022070425-appb-000038
is the final set of matching point pairs, and γ is the custom threshold.
得到
Figure PCTCN2022070425-appb-000039
后,通过下式得到上一级的搜索范围集合,
get
Figure PCTCN2022070425-appb-000039
After that, the upper-level search range set is obtained by the following formula,
Figure PCTCN2022070425-appb-000040
Figure PCTCN2022070425-appb-000040
Figure PCTCN2022070425-appb-000041
Figure PCTCN2022070425-appb-000041
其中,
Figure PCTCN2022070425-appb-000042
Figure PCTCN2022070425-appb-000043
的数量,
Figure PCTCN2022070425-appb-000044
为第l-1级相对于第l级的神经网络感受野,点p的坐标为(p x,p y,p z)。以及得到上一级的搜索范围集合后,重复上述步骤,得到最终的输出结果
Figure PCTCN2022070425-appb-000045
即为两个图像的匹配点对集合。
in,
Figure PCTCN2022070425-appb-000042
for
Figure PCTCN2022070425-appb-000043
quantity,
Figure PCTCN2022070425-appb-000044
is the receptive field of the neural network of level l-1 relative to level l, and the coordinates of point p are (p x , p y , p z ). And after getting the upper-level search range set, repeat the above steps to get the final output result
Figure PCTCN2022070425-appb-000045
That is, the set of matching point pairs of the two images.
在一优选的实施方式中,基于标记点的图像配准方法还包括:通过最小化所有匹配点对间点距离的总和,拟合出刚性配准的变换矩阵和位移向量,从而得到刚性配准后的医学图像warped image,包括以下步骤:得到所有匹配点对后,通过最小化下式,得到刚性配准的变换矩阵和位移向量的最优解:In a preferred embodiment, the image registration method based on marker points further includes: by minimizing the sum of the distances between all matching point pairs, fitting the transformation matrix and displacement vector of rigid registration, thereby obtaining rigid registration The warped image of the final medical image includes the following steps: After obtaining all matching point pairs, the optimal solution of the transformation matrix and displacement vector of rigid registration is obtained by minimizing the following formula:
Figure PCTCN2022070425-appb-000046
Figure PCTCN2022070425-appb-000046
最优解为:The optimal solution is:
R=(P TP) -1P TQ R=(P T P) -1 P T Q
A=R[0:3,0:3]A=R[0:3,0:3]
b=R[0:3,3]b=R[0:3,3]
其中,N为匹配点对个数,p n为fixed image的第n个匹配点,q n为对应的moving image中的像素点。P为fixed image所有匹配点组成的矩阵,大小为[N,4],即N个四维行向量组成的矩阵,四维的前三维是像素点的物理坐标,第四维是固定值1。Q为moving image所有匹配点组成的矩阵,大小为[N,4]。矩阵R的大小为[4,4],R[0:3,0:3]指取矩阵R的前3行、前3列组成的大小 为[3,3]的矩阵,R[0:3,3]指取矩阵R的前3行第3列的三维列向量。A和b分别为变换矩阵和位移向量的最优解。以及最后通过A和b得到刚性配准后的医学图像warped image。 Among them, N is the number of matching point pairs, p n is the nth matching point of the fixed image, and q n is the pixel point in the corresponding moving image. P is a matrix composed of all matching points of the fixed image, with a size of [N,4], that is, a matrix composed of N four-dimensional row vectors. The first three dimensions of the four dimensions are the physical coordinates of the pixels, and the fourth dimension is a fixed value of 1. Q is a matrix composed of all matching points of the moving image, with a size of [N, 4]. The size of matrix R is [4,4], R[0:3,0:3] refers to the matrix of size [3,3] composed of the first 3 rows and first 3 columns of matrix R, R[0:3 ,3] refers to the three-dimensional column vector of the first 3 rows and the 3rd column of the matrix R. A and b are the optimal solution of transformation matrix and displacement vector respectively. And finally, the warped image of the medical image after rigid registration is obtained through A and b.
在一优选的实施方式中,基于标记点的图像配准方法还包括:在刚性配准的基础上,通过基于径向基的插值法得到非刚性配准的位移场三维矩阵,从而得到非刚性配准后的医学图像warped image,包括以下步骤:位移场三维矩阵的大小与fixed image相同。在得到N个匹配点对后,采用如下插值法得到位移场矩阵剩余像素点的值:In a preferred embodiment, the marker-based image registration method further includes: on the basis of rigid registration, obtain the three-dimensional displacement field matrix of non-rigid registration by interpolation method based on radial basis, so as to obtain the non-rigid registration The warped image of the medical image after registration includes the following steps: the size of the three-dimensional matrix of the displacement field is the same as that of the fixed image. After obtaining N matching point pairs, use the following interpolation method to obtain the value of the remaining pixels of the displacement field matrix:
Figure PCTCN2022070425-appb-000047
Figure PCTCN2022070425-appb-000047
A=(a 1,a 2,a 3) A=(a 1 ,a 2 ,a 3 )
G(r)=r 2lnr G(r)=r 2 lnr
其中p为位移场三维矩阵中的坐标为(x p,y p,z p)的像素,p n为fixed image中第n个匹配点。G()为径向基函数。A、b、w n的值采用以下方式求解: Where p is the pixel whose coordinates are (x p , y p , z p ) in the three-dimensional matrix of the displacement field, and p n is the nth matching point in the fixed image. G() is the radial basis function. The values of A, b, w n are solved in the following way:
设:Assume:
Figure PCTCN2022070425-appb-000048
Figure PCTCN2022070425-appb-000048
r ij=‖p i-p jr ij =‖ p i -p j
V=(v 1,v 2,…,v N,0,0,0,0) V=(v 1 ,v 2 ,...,v N ,0,0,0,0)
v n=(q n-p n)[k]k∈(0,1,2) v n =(q n -p n )[k]k∈(0,1,2)
Figure PCTCN2022070425-appb-000049
Figure PCTCN2022070425-appb-000049
Ω=(w 1,w 2,…,w N,b,a 1,a 2,a 3) Ω=(w 1 ,w 2 ,...,w N ,b,a 1 ,a 2 ,a 3 )
其中P为fixed image所有匹配点组成的矩阵,大小为[N,4],即n个四维行向量组成的矩阵,四维的后三维是像素点的物理坐标,第一维是固定值1。q n为p n的在moving image中对应的匹配点。k为表示维度,即由于位移值是三维向量(x、y、z方向),上述求解过程只针对其中一个维度,故k=0时取x轴,故k=1时取y轴,故k=2时取z轴。 Among them, P is a matrix composed of all matching points of the fixed image, with a size of [N, 4], that is, a matrix composed of n four-dimensional row vectors. The last three dimensions of the four dimensions are the physical coordinates of the pixels, and the first dimension is a fixed value of 1. q n is the corresponding matching point of p n in the moving image. k is the dimension, that is, because the displacement value is a three-dimensional vector (x, y, z direction), the above solution process is only for one of the dimensions, so when k=0, the x axis is taken, so when k=1, the y axis is taken, so k = 2, take the z axis.
由v n=f(p n)有: From v n =f(p n ), we have:
V=LΩ T V= LΩT
进而求得所有待解参数的值:And then find the value of all parameters to be solved:
Ω=(L -1V) T Ω=(L - 1V) T
位移场中除匹配点外的像素点的位移值均可通过f()拟合得到。由于位移值是三维向量,即x、y、z方向,故上述插值过程需重复3次,即每个方向分别进行一次。以及最后得到位移场三维矩阵,从而得到非刚性配准后的医学图像warped image。The displacement values of the pixel points in the displacement field except the matching points can be obtained by f() fitting. Since the displacement value is a three-dimensional vector, that is, the x, y, and z directions, the above interpolation process needs to be repeated three times, that is, each direction is performed once. And finally obtain the three-dimensional matrix of the displacement field, so as to obtain the warped image of the medical image after non-rigid registration.
与现有技术相比,本发明的基于标记点的图像配准方法具有以下有益效果:(1)本发明能对任意两个模态的图像进行配准。(2)本发明采用一个预训练的神经网络提取图像特征,该网络的训练过程包含多种不同的任务并涉及多种不同的输入模态,能有效提高特征的有效性与通用性。(3)本发明利用一个预训练的神经网络提取图像特征,通过搜索、筛选、匹配等过程得到两个图像间的代表某种语义的多个匹配点对,能有效解决标记点金标准缺乏的问题。(4)本发明通过最小化所有匹配点对间点距离的总和,求解变换矩阵和位移向量,实现刚性配准。(5)本发明通过基于径向基的插值法求解位移场三维矩阵,实现非刚性配准。Compared with the prior art, the marker-based image registration method of the present invention has the following beneficial effects: (1) The present invention can register images of any two modalities. (2) The present invention adopts a pre-trained neural network to extract image features. The training process of this network includes multiple different tasks and involves multiple different input modes, which can effectively improve the effectiveness and versatility of features. (3) The present invention utilizes a pre-trained neural network to extract image features, and obtains multiple matching point pairs representing certain semantics between two images through processes such as search, screening, and matching, which can effectively solve the problem of lack of gold standard for marking points question. (4) The present invention realizes rigid registration by minimizing the sum of the distances between all matching point pairs and solving the transformation matrix and displacement vector. (5) The present invention solves the three-dimensional matrix of the displacement field through the interpolation method based on the radial basis to realize non-rigid registration.
附图说明Description of drawings
图1是根据本发明一实施方式的基于标记点的图像配准方法的流程示意图。Fig. 1 is a schematic flow chart of an image registration method based on markers according to an embodiment of the present invention.
具体实施方式Detailed ways
下面结合附图,对本发明的具体实施方式进行详细描述,但应当理解本发明的保护范围并不受具体实施方式的限制。The specific embodiments of the present invention will be described in detail below in conjunction with the accompanying drawings, but it should be understood that the protection scope of the present invention is not limited by the specific embodiments.
除非另有其它明确表示,否则在整个说明书和权利要求书中,术语“包括”或其变换如“包含”或“包括有”等等将被理解为包括所陈述的元件或组成部分,而并未排除其它元件或其它组成部分。Unless expressly stated otherwise, throughout the specification and claims, the term "comprise" or variations thereof such as "includes" or "includes" and the like will be understood to include the stated elements or constituents, and not Other elements or other components are not excluded.
如图1所示,根据本发明优选实施方式的一种基于标记点的图像配准方法,包括以下步骤:As shown in Figure 1, a kind of image registration method based on marker point according to the preferred embodiment of the present invention, comprises the following steps:
输入两个任意模态(CT、CBCT、MRI、PET等)的医学图像,一个作为fixed image,另一个作为moving image;Input two medical images of any modality (CT, CBCT, MRI, PET, etc.), one as a fixed image and the other as a moving image;
采用一个预训练的神经网络提取两个输入图像的金字塔特征,该网络的训练过程包含多种不同的任务并涉及上述多种不同的输入模态;Pyramid features of two input images are extracted using a pre-trained neural network, which is trained on a variety of tasks and involves the various input modalities mentioned above;
利用上述神经网络提取金字塔特征,通过搜索、筛选、匹配等过程得到两个图像间的代表某种语义的多个匹配点对;Use the above neural network to extract pyramid features, and obtain multiple matching point pairs representing certain semantics between two images through searching, screening, matching and other processes;
通过最小化所有匹配点对间点距离的总和,拟合出刚性配准的变换矩阵和位移向量,从而得到刚性配准后的医学图像warped image。By minimizing the sum of the distances between all matching point pairs, the transformation matrix and displacement vector for rigid registration are fitted to obtain the warped image of the medical image after rigid registration.
在刚性配准的基础上,通过基于径向基的插值法得到非刚性配准的位移场三维矩阵,从而得到非刚性配准后的医学图像warped image。On the basis of rigid registration, the three-dimensional matrix of the displacement field of non-rigid registration is obtained by interpolation method based on radial basis, so as to obtain the warped image of the medical image after non-rigid registration.
本发明的一种基于标记点的图像配准方法的一个具体实施的工作流程包括:A specific implementation workflow of a marker-based image registration method of the present invention includes:
在一些实施方式中,步骤S1、构建一个预训练的神经网络提取两个输入图像的金字塔特征;In some embodiments, step S1, constructing a pre-trained neural network to extract the pyramid features of two input images;
步骤S1具体包括以下步骤:Step S1 specifically includes the following steps:
S11、神经网络的结构分为骨干网络和后续多个分支网络。骨干网络在不同任务间共享,每个分支网络对应一个任务。最后用于提取图像特征的是骨干网络。S11. The structure of the neural network is divided into a backbone network and a plurality of subsequent branch networks. The backbone network is shared among different tasks, and each branch network corresponds to a task. Finally, the backbone network is used to extract image features.
S12、神经网络的训练过程包含多种不同的任务并涉及多种不同的输入模态,包括但不限于:基于CT的鼻咽癌原发肿瘤(GTV)分割、基于MRI的鼻咽癌原发肿瘤分割、基于CT的子宫颈癌原发肿瘤分割、基于PET的肺部原发肿瘤分割、基于CT的危及器官(OAR)分割、基于MRI的危及器官分割、基于CBCT的危及器官分割、基于CT的肺结节目标检测等。S12. The training process of the neural network involves a variety of different tasks and involves a variety of different input modalities, including but not limited to: CT-based nasopharyngeal carcinoma primary tumor (GTV) segmentation, MRI-based nasopharyngeal carcinoma primary Tumor segmentation, CT-based cervical cancer primary tumor segmentation, PET-based lung primary tumor segmentation, CT-based organ-at-risk (OAR) segmentation, MRI-based organ-at-risk segmentation, CBCT-based organ-at-risk segmentation, CT-based Pulmonary nodule target detection, etc.
S13、先用其中一个任务训练神经网络,再加上其它输入模态的任务各一个同时训练,再分别对剩余的每个任务单独进行训练,训练时骨干网络参数固定,最后所有任务同时训练微调所有参数。S13. First use one of the tasks to train the neural network, and then train each of the other input modal tasks at the same time, and then train each of the remaining tasks separately. The parameters of the backbone network are fixed during training, and finally all tasks are trained and fine-tuned at the same time. All parameters.
在一些实施方式中,基于标记点的图像配准方法还包括:In some embodiments, the image registration method based on markers also includes:
S2、利用上述神经网络提取金字塔特征,通过搜索、筛选、匹配等过程得到两个图像间的代表某种语义的多个匹配点对;S2. Utilize the above-mentioned neural network to extract pyramid features, and obtain multiple matching point pairs representing certain semantics between the two images through processes such as searching, screening, and matching;
步骤S2具体包括以下步骤:Step S2 specifically includes the following steps:
S21、把I f(fixed image)与I m(moving image)输入上述预训练的神经网络,提取两个输入图像的金字塔特征图(feature map)
Figure PCTCN2022070425-appb-000050
Figure PCTCN2022070425-appb-000051
其中l∈{1,2,3,4,5}表示第l级特征,数字越大表示层数越深,即特征尺寸越小但蕴含更多高层语义。
S21, input I f (fixed image) and I m (moving image) into the above-mentioned pre-trained neural network, and extract pyramidal feature maps (feature maps) of two input images
Figure PCTCN2022070425-appb-000050
and
Figure PCTCN2022070425-appb-000051
Among them, l∈{1,2,3,4,5} represents the l-level feature, and the larger the number, the deeper the layer, that is, the smaller the feature size is, the more high-level semantics it contains.
S22、匹配点的搜索需要在特定的搜索范围内产生,从l=5开始设置搜索范围:S22, the search for matching points needs to be generated within a specific search range, and the search range is set from l=5:
S 5={(P 5,Q 5)} S 5 ={(P 5 ,Q 5 )}
其中:in:
Figure PCTCN2022070425-appb-000052
Figure PCTCN2022070425-appb-000052
Figure PCTCN2022070425-appb-000053
为I f的第l级的第n个搜索范围,对应地
Figure PCTCN2022070425-appb-000054
为I m的第l级的第n个搜索范围,N l为第l级的搜索范围的个数,S l为第l级的多个搜索范围对的集合。当l=5时,搜索范围是
Figure PCTCN2022070425-appb-000055
Figure PCTCN2022070425-appb-000056
的整个范围,即N 5=1。
Figure PCTCN2022070425-appb-000053
is the nth search range of the lth level of If , correspondingly
Figure PCTCN2022070425-appb-000054
is the nth search range of the lth level of Im , N1 is the number of the lth level search range, and S1 is a set of multiple search range pairs of the lth level. When l=5, the search range is
Figure PCTCN2022070425-appb-000055
and
Figure PCTCN2022070425-appb-000056
The entire range of , that is, N 5 =1.
S23、通过下式对搜索范围
Figure PCTCN2022070425-appb-000057
Figure PCTCN2022070425-appb-000058
的特征图
Figure PCTCN2022070425-appb-000059
Figure PCTCN2022070425-appb-000060
进行变换,得到
Figure PCTCN2022070425-appb-000061
Figure PCTCN2022070425-appb-000062
S23, through the following formula to search range
Figure PCTCN2022070425-appb-000057
and
Figure PCTCN2022070425-appb-000058
feature map of
Figure PCTCN2022070425-appb-000059
and
Figure PCTCN2022070425-appb-000060
transform to get
Figure PCTCN2022070425-appb-000061
and
Figure PCTCN2022070425-appb-000062
Figure PCTCN2022070425-appb-000063
Figure PCTCN2022070425-appb-000063
Figure PCTCN2022070425-appb-000064
Figure PCTCN2022070425-appb-000064
Figure PCTCN2022070425-appb-000065
Figure PCTCN2022070425-appb-000065
Figure PCTCN2022070425-appb-000066
Figure PCTCN2022070425-appb-000066
其中,
Figure PCTCN2022070425-appb-000067
表示在
Figure PCTCN2022070425-appb-000068
范围内的局部特征图,
Figure PCTCN2022070425-appb-000069
为变换后的特征图,
Figure PCTCN2022070425-appb-000070
Figure PCTCN2022070425-appb-000071
的均值,
Figure PCTCN2022070425-appb-000072
Figure PCTCN2022070425-appb-000073
的标准差,同
Figure PCTCN2022070425-appb-000074
Figure PCTCN2022070425-appb-000075
in,
Figure PCTCN2022070425-appb-000067
expressed in
Figure PCTCN2022070425-appb-000068
Local feature maps in the range,
Figure PCTCN2022070425-appb-000069
is the transformed feature map,
Figure PCTCN2022070425-appb-000070
for
Figure PCTCN2022070425-appb-000071
the mean value of
Figure PCTCN2022070425-appb-000072
for
Figure PCTCN2022070425-appb-000073
standard deviation of
Figure PCTCN2022070425-appb-000074
Figure PCTCN2022070425-appb-000075
S24、在搜索范围
Figure PCTCN2022070425-appb-000076
Figure PCTCN2022070425-appb-000077
内搜索匹配点对,当满足以下条件时,两个点p l与q l为匹配点对:
S24. In the search range
Figure PCTCN2022070425-appb-000076
and
Figure PCTCN2022070425-appb-000077
Search for matching point pairs within, when the following conditions are met, the two points p l and q l are matching point pairs:
Figure PCTCN2022070425-appb-000078
Figure PCTCN2022070425-appb-000078
Figure PCTCN2022070425-appb-000079
Figure PCTCN2022070425-appb-000079
Figure PCTCN2022070425-appb-000080
Figure PCTCN2022070425-appb-000080
即若以
Figure PCTCN2022070425-appb-000081
内的点p l
Figure PCTCN2022070425-appb-000082
范围内搜索时,相似度最高的点是q l,反之也成立,则p l与q l为匹配点对。相似度的计算公式为:
That is to say, if
Figure PCTCN2022070425-appb-000081
The point p l inside
Figure PCTCN2022070425-appb-000082
When searching within the range, the point with the highest similarity is q l , and vice versa, then p l and q l are matching point pairs. The formula for calculating the similarity is:
Figure PCTCN2022070425-appb-000083
Figure PCTCN2022070425-appb-000083
其中,ε(p l)为点p l的特定范围内邻域的点集。 Among them, ε(p l ) is the point set of the neighborhood within a specific range of point p l .
S25、第l级的所有N l个搜索范围分别执行上述搜索匹配点的步骤S23~S24,便得到所有匹配点对的集合Λ lS25, all the N1 search ranges of the first level perform the above steps S23-S24 of searching for matching points respectively, and then obtain the set Λl of all matching point pairs:
Figure PCTCN2022070425-appb-000084
Figure PCTCN2022070425-appb-000084
S26、对于上述步骤搜索得到的匹配点对,还必须通过以下的筛选条件,即该点在特征图(feature map)的值必须足够大:S26. For the matching point pair obtained by searching in the above steps, the following screening conditions must also be passed, that is, the value of the point in the feature map (feature map) must be large enough:
Figure PCTCN2022070425-appb-000085
Figure PCTCN2022070425-appb-000085
Figure PCTCN2022070425-appb-000086
Figure PCTCN2022070425-appb-000086
其中,
Figure PCTCN2022070425-appb-000087
为最终得到的匹配点对集合,γ为自定义阈值,这里取0.05。
in,
Figure PCTCN2022070425-appb-000087
is the final set of matching point pairs, and γ is a custom threshold, which is 0.05 here.
S27、得到
Figure PCTCN2022070425-appb-000088
后,通过下式得到上一级的搜索范围集合,
S27, get
Figure PCTCN2022070425-appb-000088
After that, the upper-level search range set is obtained by the following formula,
Figure PCTCN2022070425-appb-000089
Figure PCTCN2022070425-appb-000089
Figure PCTCN2022070425-appb-000090
Figure PCTCN2022070425-appb-000090
其中,
Figure PCTCN2022070425-appb-000091
Figure PCTCN2022070425-appb-000092
的数量,
Figure PCTCN2022070425-appb-000093
为第l-1级相对于第l级的神经网络感受野,点p的坐标为(p x,p y,p z)。
in,
Figure PCTCN2022070425-appb-000091
for
Figure PCTCN2022070425-appb-000092
quantity,
Figure PCTCN2022070425-appb-000093
is the receptive field of the neural network of level l-1 relative to level l, and the coordinates of point p are (p x , p y , p z ).
S28、得到上一级的搜索范围集合后,对每个搜索范围,跳转至步骤S23并重复步骤S23~S28,得到最终的输出结果
Figure PCTCN2022070425-appb-000094
即为两个图像的匹配点对集合。
S28. After obtaining the upper-level search range set, for each search range, jump to step S23 and repeat steps S23-S28 to obtain the final output result
Figure PCTCN2022070425-appb-000094
That is, the set of matching point pairs of the two images.
在一些实施方式中,基于标记点的图像配准方法还包括:In some embodiments, the image registration method based on markers also includes:
S3、通过最小化所有匹配点对间点距离的总和,拟合出刚性配准的变换矩阵和位移向量,从而得到刚性配准后的医学图像warped image;S3, by minimizing the sum of the distances between all matching point pairs, fitting the transformation matrix and displacement vector of the rigid registration, thereby obtaining the warped image of the medical image after the rigid registration;
步骤S3具体包括以下步骤:Step S3 specifically includes the following steps:
S31、得到所有匹配点对后,通过最小化下式,得到刚性配准的变换矩阵和位移向量的最优解:S31. After all matching point pairs are obtained, the optimal solution of the transformation matrix and displacement vector for rigid registration is obtained by minimizing the following formula:
Figure PCTCN2022070425-appb-000095
Figure PCTCN2022070425-appb-000095
最优解为:The optimal solution is:
R=(P TP) -1P TQ R=(P T P) -1 P T Q
A=R[0:3,0:3]A=R[0:3,0:3]
b=R[0:3,3]b=R[0:3,3]
其中,N为匹配点对个数,p n为fixed image的第n个匹配点,q n为对应的moving image中的像素点。P为fixed image所有匹配点组成的矩阵,大小为[N,4],即N个四维行向量组成的矩阵,四维的前三维是像素点的物理坐标,第四维是固定值1。Q为moving image所有匹配点组成的矩阵,大小为[N,4]。矩阵R的大小为[4,4],R[0:3,0:3]指取矩阵R的前3行、前3列组成的大小为[3,3]的矩阵,R[0:3,3]指取矩阵R的前3行第3列的三维列向量。A和b分别为变换矩阵和位移向量的最优解。 Among them, N is the number of matching point pairs, p n is the nth matching point of the fixed image, and q n is the pixel point in the corresponding moving image. P is a matrix composed of all matching points of the fixed image, with a size of [N,4], that is, a matrix composed of N four-dimensional row vectors. The first three dimensions of the four dimensions are the physical coordinates of the pixels, and the fourth dimension is a fixed value of 1. Q is a matrix composed of all matching points of the moving image, with a size of [N, 4]. The size of matrix R is [4,4], R[0:3,0:3] refers to the matrix of size [3,3] composed of the first 3 rows and first 3 columns of matrix R, R[0:3 ,3] refers to the three-dimensional column vector of the first 3 rows and the 3rd column of the matrix R. A and b are the optimal solution of transformation matrix and displacement vector respectively.
S32、最后通过A和b得到刚性配准后的医学图像warped image。S32. Finally, the warped image of the medical image after rigid registration is obtained through A and b.
在一些实施方式中,基于标记点的图像配准方法还包括:In some embodiments, the image registration method based on markers also includes:
S4、在刚性配准的基础上,通过基于径向基的插值法得到非刚性配准的位移场三维矩阵,从而得到非刚性配准后的医学图像warped image;S4. On the basis of rigid registration, the three-dimensional matrix of the displacement field of non-rigid registration is obtained by interpolation method based on radial basis, so as to obtain the warped image of the medical image after non-rigid registration;
步骤S4具体包括以下步骤:Step S4 specifically includes the following steps:
S41、位移场三维矩阵的大小与fixed image相同。在得到N个匹配点对后,采用如下插值法得到位移场矩阵剩余像素点的值:S41. The size of the displacement field three-dimensional matrix is the same as that of the fixed image. After obtaining N matching point pairs, use the following interpolation method to obtain the value of the remaining pixels of the displacement field matrix:
Figure PCTCN2022070425-appb-000096
Figure PCTCN2022070425-appb-000096
A=(a 1,a 2,a 3) A=(a 1 ,a 2 ,a 3 )
G(r)=r 2lnr G(r)=r 2 lnr
其中p为位移场三维矩阵中的坐标为(x p,y p,z p)的像素,p n为fixed image 中第n个匹配点。G()为径向基函数。A、b、w n的值采用以下方式求解: Where p is the pixel whose coordinates are (x p , y p , z p ) in the three-dimensional matrix of the displacement field, and p n is the nth matching point in the fixed image. G() is the radial basis function. The values of A, b, w n are solved in the following way:
设:Assume:
Figure PCTCN2022070425-appb-000097
Figure PCTCN2022070425-appb-000097
r ij=‖p i-p jr ij =‖ p i -p j
V=(v 1,v 2,…,v N,0,0,0,0) V=(v 1 ,v 2 ,...,v N ,0,0,0,0)
v n=(q n-p n)[k]k∈(0,1,2) v n =(q n -p n )[k]k∈(0,1,2)
Figure PCTCN2022070425-appb-000098
Figure PCTCN2022070425-appb-000098
Ω=(w 1,w 2,…,w N,b,a 1,a 2,a 3) Ω=(w 1 ,w 2 ,...,w N ,b,a 1 ,a 2 ,a 3 )
其中P为fixed image所有匹配点组成的矩阵,大小为[N,4],即n个四维行向量组成的矩阵,四维的后三维是像素点的物理坐标,第一维是固定值1。q n为p n的在moving image中对应的匹配点。k为表示维度,即由于位移值是三维向量(x、y、z方向),上述求解过程只针对其中一个维度,故k=0时取x轴,故k=1时取y轴,故k=2时取z轴。 Among them, P is a matrix composed of all matching points of the fixed image, with a size of [N, 4], that is, a matrix composed of n four-dimensional row vectors. The last three dimensions of the four dimensions are the physical coordinates of the pixels, and the first dimension is a fixed value of 1. q n is the corresponding matching point of p n in the moving image. k is the dimension, that is, because the displacement value is a three-dimensional vector (x, y, z direction), the above solution process is only for one of the dimensions, so when k=0, the x axis is taken, so when k=1, the y axis is taken, so k = 2, take the z axis.
由v n=f(p n)有: From v n =f(p n ), we have:
V=LΩ T V= LΩT
进而求得所有待解参数的值:And then find the value of all parameters to be solved:
Ω=(L -1V) T Ω=(L - 1V) T
位移场中除匹配点外的像素点的位移值均可通过f()拟合得到。The displacement values of the pixel points in the displacement field except the matching points can be obtained by f() fitting.
S42、由于位移值是三维向量,即x、y、z方向,故上述插值过程需重复3次,即每个方向分别进行一次步骤S41。S42. Since the displacement value is a three-dimensional vector, that is, the x, y, and z directions, the above interpolation process needs to be repeated three times, that is, step S41 is performed once for each direction.
S43、最后得到位移场三维矩阵,从而得到非刚性配准后的医学图像warped image。S43. Finally, the three-dimensional matrix of the displacement field is obtained, so as to obtain the warped image of the medical image after non-rigid registration.
综上所述,本发明的基于标记点的图像配准方法具有以下优点:(1)本发明能对任意两个模态的图像进行配准。(2)本发明采用一个预训练的神经网络提取图像特征,该网络的训练过程包含多种不同的任务并涉及多种不同 的输入模态,能有效提高特征的有效性与通用性。(3)本发明利用一个预训练的神经网络提取图像特征,通过搜索、筛选、匹配等过程得到两个图像间的代表某种语义的多个匹配点对,能有效解决标记点金标准缺乏的问题。(4)本发明通过最小化所有匹配点对间点距离的总和,求解变换矩阵和位移向量,实现刚性配准。(5)本发明通过基于径向基的插值法求解位移场三维矩阵,实现非刚性配准。In summary, the marker-based image registration method of the present invention has the following advantages: (1) The present invention can register images of any two modalities. (2) The present invention adopts a pre-trained neural network to extract image features, and the training process of this network includes multiple different tasks and involves multiple different input modes, which can effectively improve the effectiveness and versatility of features. (3) The present invention utilizes a pre-trained neural network to extract image features, and obtains multiple matching point pairs representing certain semantics between two images through processes such as search, screening, and matching, which can effectively solve the problem of lack of gold standard for marking points question. (4) The present invention realizes rigid registration by minimizing the sum of the distances between all matching point pairs and solving the transformation matrix and displacement vector. (5) The present invention solves the three-dimensional matrix of the displacement field through the interpolation method based on the radial basis to realize non-rigid registration.
前述对本发明的具体示例性实施方案的描述是为了说明和例证的目的。这些描述并非想将本发明限定为所公开的精确形式,并且很显然,根据上述教导,可以进行很多改变和变化。对示例性实施例进行选择和描述的目的在于解释本发明的特定原理及其实际应用,从而使得本领域的技术人员能够实现并利用本发明的各种不同的示例性实施方案以及各种不同的选择和改变。本发明的范围意在由权利要求书及其等同形式所限定。The foregoing descriptions of specific exemplary embodiments of the present invention have been presented for purposes of illustration and description. These descriptions are not intended to limit the invention to the precise form disclosed, and obviously many modifications and variations are possible in light of the above teaching. The exemplary embodiments were chosen and described in order to explain the specific principles of the invention and its practical application, thereby enabling others skilled in the art to make and use various exemplary embodiments of the invention, as well as various Choose and change. It is intended that the scope of the invention be defined by the claims and their equivalents.

Claims (5)

  1. 一种基于标记点的图像配准方法,其特征在于,包括以下步骤:A method for image registration based on markers, comprising the following steps:
    输入两个任意模态的医学图像,一个作为fixed image,另一个作为moving image;Input two medical images of any modality, one as fixed image and the other as moving image;
    采用一个预训练的神经网络提取两个输入任意模态的所述医学图像的金字塔特征,所述神经网络的训练过程包含多种任务并涉及多种输入模态;Adopting a pre-trained neural network to extract the pyramidal features of the medical image of two input arbitrary modalities, the training process of the neural network includes multiple tasks and involves multiple input modalities;
    利用所述神经网络提取的所述金字塔特征,通过搜索、筛选及匹配过程得到两个图像间的代表某种语义的多个匹配点对;Using the pyramid features extracted by the neural network to obtain a plurality of matching point pairs representing certain semantics between the two images through searching, screening and matching processes;
    通过最小化所有匹配点对间点距离的总和,拟合出刚性配准的变换矩阵和位移向量,从而得到刚性配准后的医学图像warped image;以及By minimizing the sum of the distances between all matching point pairs, the transformation matrix and displacement vector of rigid registration are fitted, so as to obtain the warped image of the medical image after rigid registration; and
    在刚性配准的基础上,通过基于径向基的插值法得到非刚性配准的位移场三维矩阵,从而得到非刚性配准后的医学图像warped image;On the basis of rigid registration, the three-dimensional matrix of displacement field of non-rigid registration is obtained by interpolation method based on radial basis, so as to obtain the medical image warped image after non-rigid registration;
    其中所述采用一个预训练的所述神经网络提取两个输入的任意模态的所述医学图像的所述金字塔特征,包括:所述神经网络的结构分为骨干网络和后续多个分支网络;所述骨干网络在不同任务间共享,每个所述分支网络对应一个任务;最后用于提取图像特征的是所述骨干网络;以及Wherein said using a pre-trained neural network to extract the pyramidal features of the medical images of two input arbitrary modalities includes: the structure of the neural network is divided into a backbone network and a plurality of subsequent branch networks; The backbone network is shared between different tasks, and each of the branch networks corresponds to a task; the backbone network is finally used to extract image features; and
    先用其中一个任务训练的所述神经网络,再加上其它输入模态的任务各一个同时训练,再分别对剩余的每个任务单独进行训练,训练时所述骨干网络参数固定,最后所有任务同时训练微调所有参数。First use the neural network trained by one of the tasks, and then train each of the other input modal tasks at the same time, and then train the remaining tasks separately. During training, the backbone network parameters are fixed, and finally all tasks Simultaneously train and fine-tune all parameters.
  2. 根据权利要求1所述的基于标记点的图像配准方法,其特征在于,所述采用一个预训练的所述神经网络提取两个输入的任意模态的所述医学图像的所述金字塔特征,还包括:The image registration method based on markers according to claim 1, wherein the pyramid feature of the medical image of the arbitrary modality of the two input is extracted by using a pre-trained neural network, Also includes:
    所述神经网络的训练过程包含多种不同的任务并涉及多种不同的输入模态,包括:基于CT的鼻咽癌原发肿瘤分割、基于MRI的鼻咽癌原发肿瘤分割、基于CT的子宫颈癌原发肿瘤分割、基于PET的肺部原发肿瘤分割、基于CT的危及器官分割、基于MRI的危及器官分割、基于CBCT的危及器官 分割以及基于CT的肺结节目标检测。The training process of the neural network includes a variety of different tasks and involves a variety of different input modalities, including: CT-based nasopharyngeal carcinoma primary tumor segmentation, MRI-based nasopharyngeal carcinoma primary tumor segmentation, CT-based Cervical cancer primary tumor segmentation, PET-based lung primary tumor segmentation, CT-based organ-at-risk segmentation, MRI-based organ-at-risk segmentation, CBCT-based organ-at-risk segmentation, and CT-based pulmonary nodule target detection.
  3. 根据权利要求1所述的基于标记点的图像配准方法,其特征在于,利用所述神经网络提取的所述金字塔特征,通过搜索、筛选及匹配过程得到两个图像间的代表某种语义的多个匹配点对,包括以下步骤:The image registration method based on markers according to claim 1, wherein the pyramid feature extracted by the neural network is used to obtain the representation of certain semantics between the two images through searching, screening and matching processes. Multiple matching point pairs, including the following steps:
    把I f与I m输入上述预训练的神经网络,提取两个输入图像的金字塔特征图的
    Figure PCTCN2022070425-appb-100001
    Figure PCTCN2022070425-appb-100002
    其中l∈{1,2,3,4,5}表示第l级特征,数字越大表示层数越深,即特征尺寸越小但蕴含更多高层语义;其中I f即为fixed image,I m即为moving image;
    Input If and Im into the above-mentioned pre-trained neural network, and extract the pyramidal feature maps of the two input images
    Figure PCTCN2022070425-appb-100001
    and
    Figure PCTCN2022070425-appb-100002
    Among them, l∈{1,2,3,4,5} represents the l-level feature, and the larger the number, the deeper the layer, that is, the smaller the feature size but contains more high-level semantics; where I f is the fixed image, I m is the moving image;
    匹配点的搜索需要在特定的搜索范围内产生,从l=5开始设置搜索范围:The search for matching points needs to be generated within a specific search range, and the search range is set starting from l=5:
    S 5={(P 5,Q 5)} S 5 ={(P 5 ,Q 5 )}
    其中:in:
    Figure PCTCN2022070425-appb-100003
    Figure PCTCN2022070425-appb-100003
    Figure PCTCN2022070425-appb-100004
    为I f的第l级的第n个搜索范围,对应地
    Figure PCTCN2022070425-appb-100005
    为I m的第l级的第n个搜索范围,N l为第l级的搜索范围的个数,S l为第l级的多个搜索范围对的集合;当l=5时,搜索范围是
    Figure PCTCN2022070425-appb-100006
    Figure PCTCN2022070425-appb-100007
    的整个范围,即N 5=1;
    Figure PCTCN2022070425-appb-100004
    is the nth search range of the lth level of If , correspondingly
    Figure PCTCN2022070425-appb-100005
    is the nth search range of the lth level of I m , N1 is the number of the search range of the lth level, and S1 is the collection of multiple search range pairs of the lth level; when l=5, the search range Yes
    Figure PCTCN2022070425-appb-100006
    and
    Figure PCTCN2022070425-appb-100007
    The entire range of , that is, N 5 =1;
    通过下式对搜索范围
    Figure PCTCN2022070425-appb-100008
    Figure PCTCN2022070425-appb-100009
    的特征图
    Figure PCTCN2022070425-appb-100010
    Figure PCTCN2022070425-appb-100011
    进行变换,得到
    Figure PCTCN2022070425-appb-100012
    Figure PCTCN2022070425-appb-100013
    Search range by the following formula
    Figure PCTCN2022070425-appb-100008
    and
    Figure PCTCN2022070425-appb-100009
    feature map of
    Figure PCTCN2022070425-appb-100010
    and
    Figure PCTCN2022070425-appb-100011
    transform to get
    Figure PCTCN2022070425-appb-100012
    and
    Figure PCTCN2022070425-appb-100013
    Figure PCTCN2022070425-appb-100014
    Figure PCTCN2022070425-appb-100014
    Figure PCTCN2022070425-appb-100015
    Figure PCTCN2022070425-appb-100015
    Figure PCTCN2022070425-appb-100016
    Figure PCTCN2022070425-appb-100016
    Figure PCTCN2022070425-appb-100017
    Figure PCTCN2022070425-appb-100017
    其中,
    Figure PCTCN2022070425-appb-100018
    表示在
    Figure PCTCN2022070425-appb-100019
    范围内的局部特征图,
    Figure PCTCN2022070425-appb-100020
    为变换后的特征图,
    Figure PCTCN2022070425-appb-100021
    Figure PCTCN2022070425-appb-100022
    的均值,
    Figure PCTCN2022070425-appb-100023
    Figure PCTCN2022070425-appb-100024
    的标准差,同
    Figure PCTCN2022070425-appb-100025
    Figure PCTCN2022070425-appb-100026
    in,
    Figure PCTCN2022070425-appb-100018
    expressed in
    Figure PCTCN2022070425-appb-100019
    Local feature maps in the range,
    Figure PCTCN2022070425-appb-100020
    is the transformed feature map,
    Figure PCTCN2022070425-appb-100021
    for
    Figure PCTCN2022070425-appb-100022
    the mean value of
    Figure PCTCN2022070425-appb-100023
    for
    Figure PCTCN2022070425-appb-100024
    standard deviation of
    Figure PCTCN2022070425-appb-100025
    Figure PCTCN2022070425-appb-100026
    在搜索范围
    Figure PCTCN2022070425-appb-100027
    Figure PCTCN2022070425-appb-100028
    内搜索匹配点对,当满足以下条件时,两个点p l与q l为匹配点对:
    in the search range
    Figure PCTCN2022070425-appb-100027
    and
    Figure PCTCN2022070425-appb-100028
    Search for matching point pairs within, when the following conditions are met, the two points p l and q l are matching point pairs:
    Figure PCTCN2022070425-appb-100029
    Figure PCTCN2022070425-appb-100029
    Figure PCTCN2022070425-appb-100030
    Figure PCTCN2022070425-appb-100030
    Figure PCTCN2022070425-appb-100031
    Figure PCTCN2022070425-appb-100031
    即若以
    Figure PCTCN2022070425-appb-100032
    内的点p l
    Figure PCTCN2022070425-appb-100033
    范围内搜索时,相似度最高的点是q l,反之也成立,则p l与q l为匹配点对;相似度的计算公式为:
    That is to say, if
    Figure PCTCN2022070425-appb-100032
    The point p l inside
    Figure PCTCN2022070425-appb-100033
    When searching within the range, the point with the highest similarity is q l , and vice versa, then p l and q l are matching point pairs; the calculation formula of similarity is:
    Figure PCTCN2022070425-appb-100034
    Figure PCTCN2022070425-appb-100034
    其中,ε(p l)为点p l的特定范围内邻域的点集; Among them, ε(p l ) is the point set of the neighborhood within a specific range of point p l ;
    第l级的所有N l个搜索范围分别执行上述搜索匹配点的步骤,便得到所有匹配点对的集合Λ lAll the N1 search ranges of the lth level perform the above-mentioned steps of searching for matching points respectively, and then obtain the set Λl of all matching point pairs:
    Figure PCTCN2022070425-appb-100035
    Figure PCTCN2022070425-appb-100035
    对于上述步骤搜索得到的匹配点对,还必须通过以下的筛选条件,即该点在特征图的值必须足够大:For the matching point pairs searched in the above steps, the following filter conditions must also be passed, that is, the value of the point in the feature map must be large enough:
    Figure PCTCN2022070425-appb-100036
    Figure PCTCN2022070425-appb-100036
    Figure PCTCN2022070425-appb-100037
    Figure PCTCN2022070425-appb-100037
    其中,
    Figure PCTCN2022070425-appb-100038
    为最终得到的匹配点对集合,γ为自定义阈值;
    in,
    Figure PCTCN2022070425-appb-100038
    is the final set of matching point pairs, and γ is the custom threshold;
    得到
    Figure PCTCN2022070425-appb-100039
    后,通过下式得到上一级的搜索范围集合:
    get
    Figure PCTCN2022070425-appb-100039
    After that, the upper-level search range set is obtained by the following formula:
    Figure PCTCN2022070425-appb-100040
    Figure PCTCN2022070425-appb-100040
    Figure PCTCN2022070425-appb-100041
    Figure PCTCN2022070425-appb-100041
    其中,
    Figure PCTCN2022070425-appb-100042
    Figure PCTCN2022070425-appb-100043
    的数量,
    Figure PCTCN2022070425-appb-100044
    为第l-1级相对于第l级的神经网络感受野,点p的坐标为(p x,p y,p z),ε(q l)为点q l的特定范围内邻域的点集,ε(q l)与ε(p l)为对应概念,p l与q l为第l级的两张图的匹配点对;以及
    in,
    Figure PCTCN2022070425-appb-100042
    for
    Figure PCTCN2022070425-appb-100043
    quantity,
    Figure PCTCN2022070425-appb-100044
    is the neural network receptive field of level l-1 relative to level l, the coordinates of point p are (p x , p y , p z ), and ε(q l ) is a point in the neighborhood of point q l within a specific range set, ε(q l ) and ε(p l ) are the corresponding concepts, p l and q l are the matching point pairs of the two graphs of level l; and
    得到上一级的搜索范围集合后,重复上述步骤,得到最终的输出结果
    Figure PCTCN2022070425-appb-100045
    即为两个图像的匹配点对集合。
    After obtaining the search scope set of the upper level, repeat the above steps to obtain the final output result
    Figure PCTCN2022070425-appb-100045
    That is, the set of matching point pairs of the two images.
  4. 根据权利要求1所述的基于标记点的图像配准方法,其特征在于,通过最小化所有匹配点对间点距离的总和,拟合出刚性配准的变换矩阵和位移向量,从而得到刚性配准后的医学图像warped image,包括以下步骤:The image registration method based on marker points according to claim 1, characterized in that, by minimizing the sum of the distances between all matching point pairs, the transformation matrix and displacement vector of rigid registration are fitted to obtain rigid registration The warped image of the medical image after the standard includes the following steps:
    得到所有匹配点对后,通过最小化下式,得到刚性配准的变换矩阵和位移向量的最优解:After all matching point pairs are obtained, the optimal solution of the transformation matrix and displacement vector for rigid registration is obtained by minimizing the following formula:
    Figure PCTCN2022070425-appb-100046
    Figure PCTCN2022070425-appb-100046
    最优解为:The optimal solution is:
    R=(P TP) -1P TQ R=(P T P) -1 P T Q
    A=R[0:3,0:3]A=R[0:3,0:3]
    b=R[0:3,3]b=R[0:3,3]
    其中,N为匹配点对个数,p n为fixed image的第n个匹配点,q n为对应的moving image中的像素点;P为fixed image所有匹配点组成的矩阵,大小为[N,4],即N个四维行向量组成的矩阵,四维的前三维是像素点的物理坐标,第四维是固定值1;Q为moving image所有匹配点组成的矩阵,大小为[N,4];矩阵R的大小为[4,4],R[0:3,0:3]指取矩阵R的前3行、前3列组成的大小为[3,3]的矩阵,R[0:3,3]指取矩阵R的前3行第3列的三维列向量;A和b分别为变换矩阵和位移向量的最优解;以及 Among them, N is the number of matching point pairs, p n is the nth matching point of the fixed image, q n is the pixel in the corresponding moving image; P is a matrix composed of all matching points of the fixed image, and the size is [N, 4], which is a matrix composed of N four-dimensional row vectors. The first three dimensions of the four dimensions are the physical coordinates of the pixels, and the fourth dimension is a fixed value of 1; Q is a matrix composed of all matching points of the moving image, with a size of [N,4] ;The size of the matrix R is [4,4], R[0:3,0:3] refers to the matrix whose size is [3,3] composed of the first 3 rows and the first 3 columns of the matrix R, R[0: 3,3] refers to the three-dimensional column vector of the first 3 rows and the 3rd column of the matrix R; A and b are the optimal solutions of the transformation matrix and the displacement vector respectively; and
    最后通过A和b得到刚性配准后的医学图像warped image。Finally, the warped image of the medical image after rigid registration is obtained through A and b.
  5. 根据权利要求1所述的基于标记点的图像配准方法,其特征在于,在刚性配准的基础上,通过基于径向基的插值法得到非刚性配准的位移场三维矩阵,从而得到非刚性配准后的医学图像warped image,包括以下步骤:The image registration method based on marker points according to claim 1, characterized in that, on the basis of rigid registration, a three-dimensional displacement field matrix of non-rigid registration is obtained by interpolation method based on radial basis, thereby obtaining non-rigid registration. The medical image warped image after rigid registration includes the following steps:
    所述位移场三维矩阵的大小与fixed image相同;在得到N个匹配点对后,采用如下插值法得到所述位移场矩阵剩余像素点的值:The size of the three-dimensional matrix of the displacement field is the same as that of the fixed image; after obtaining N matching point pairs, the following interpolation method is used to obtain the value of the remaining pixels of the displacement field matrix:
    Figure PCTCN2022070425-appb-100047
    Figure PCTCN2022070425-appb-100047
    A=(a 1,a 2,a 3) A=(a 1 ,a 2 ,a 3 )
    G(r)=r 2ln r G(r)=r 2 ln r
    其中p为所述位移场三维矩阵中的坐标为(x p,y p,z p)的像素,p n为fixed image中第n个匹配点;G( )为径向基函数;A、b、w n的值采用以下方式求解: Where p is the pixel whose coordinates are (x p , y p , z p ) in the three-dimensional matrix of the displacement field, p n is the nth matching point in the fixed image; G( ) is the radial basis function; A, b The values of , w n are solved in the following way:
    设:Assume:
    Figure PCTCN2022070425-appb-100048
    Figure PCTCN2022070425-appb-100048
    r ij=||p i-p j|| r ij =||p i -p j ||
    V=(v 1,v 2,…,v N,0,0,0,0) V=(v 1 ,v 2 ,...,v N ,0,0,0,0)
    v n=(q n-p n)[k]k∈(0,1,2) v n =(q n -p n )[k]k∈(0,1,2)
    Figure PCTCN2022070425-appb-100049
    Figure PCTCN2022070425-appb-100049
    Ω=(w 1,w 2,…,w N,b,a 1,a 2,a 3) Ω=(w 1 ,w 2 ,...,w N ,b,a 1 ,a 2 ,a 3 )
    其中P为fixed image所有匹配点组成的矩阵,大小为[N,4],即n个四维行向量组成的矩阵,四维的后三维是像素点的物理坐标,第一维是固定值1;q n为p n的在moving image中对应的匹配点;k为表示维度,即由于位移值是三维向量(x、y、z方向),上述求解过程只针对其中一个维度,故k=0时取x轴,故k=1时取y轴,故k=2时取z轴; Among them, P is a matrix composed of all matching points of the fixed image, the size is [N, 4], that is, a matrix composed of n four-dimensional row vectors, the last three dimensions of the four dimensions are the physical coordinates of the pixels, and the first dimension is a fixed value 1; q n is the matching point corresponding to p n in the moving image; k is the dimension, that is, since the displacement value is a three-dimensional vector (x, y, z direction), the above solution process is only for one of the dimensions, so when k=0, take x-axis, so take the y-axis when k=1, so take the z-axis when k=2;
    由v n=f(p n)有: From v n =f(p n ), we have:
    V=LΩ T V= LΩT
    进而求得所有待解参数的值:And then find the value of all parameters to be solved:
    Ω=(L -1V) T Ω=(L - 1V) T
    所述位移场中除匹配点外的像素点的位移值均可通过f( )拟合得到;The displacement values of the pixel points except the matching point in the displacement field can be obtained by f( ) fitting;
    由于所述位移值是三维向量,即x、y、z方向,故上述插值过程需重复3次,即每个方向分别进行一次;以及Since the displacement value is a three-dimensional vector, i.e. in the x, y, and z directions, the above interpolation process needs to be repeated 3 times, that is, once in each direction; and
    最后得到位移场三维矩阵,从而得到非刚性配准后的医学图像warped image。Finally, the three-dimensional matrix of the displacement field is obtained, so as to obtain the warped image of the medical image after non-rigid registration.
PCT/CN2022/070425 2021-05-27 2022-01-06 Mark point-based image registration method WO2022247296A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202110585049.8A CN113298854B (en) 2021-05-27 2021-05-27 Image registration method based on mark points
CN202110585049.8 2021-05-27

Publications (1)

Publication Number Publication Date
WO2022247296A1 true WO2022247296A1 (en) 2022-12-01

Family

ID=77325584

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2022/070425 WO2022247296A1 (en) 2021-05-27 2022-01-06 Mark point-based image registration method

Country Status (2)

Country Link
CN (1) CN113298854B (en)
WO (1) WO2022247296A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117474993A (en) * 2023-10-27 2024-01-30 哈尔滨工程大学 Underwater image feature point sub-pixel position estimation method and device

Families Citing this family (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113298854B (en) * 2021-05-27 2022-02-01 广州柏视医疗科技有限公司 Image registration method based on mark points
CN113744328B (en) * 2021-11-05 2022-02-15 极限人工智能有限公司 Medical image mark point identification method and device, electronic equipment and storage medium
CN113920179B (en) * 2021-11-09 2022-04-29 广州柏视医疗科技有限公司 Mark point-based multi-vision 2D-3D image non-rigid registration method and system
CN113920178B (en) * 2021-11-09 2022-04-12 广州柏视医疗科技有限公司 Mark point-based multi-vision 2D-3D image registration method and system
CN114404039B (en) * 2021-12-30 2023-05-05 华科精准(北京)医疗科技有限公司 Tissue drift correction method and device for three-dimensional model, electronic equipment and storage medium
CN114241077B (en) * 2022-02-23 2022-07-15 南昌睿度医疗科技有限公司 CT image resolution optimization method and device

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104021547A (en) * 2014-05-17 2014-09-03 清华大学深圳研究生院 Three dimensional matching method for lung CT
US20170337682A1 (en) * 2016-05-18 2017-11-23 Siemens Healthcare Gmbh Method and System for Image Registration Using an Intelligent Artificial Agent
CN109064502A (en) * 2018-07-11 2018-12-21 西北工业大学 The multi-source image method for registering combined based on deep learning and artificial design features
CN111091589A (en) * 2019-11-25 2020-05-01 北京理工大学 Ultrasonic and nuclear magnetic image registration method and device based on multi-scale supervised learning
CN111931929A (en) * 2020-07-29 2020-11-13 深圳地平线机器人科技有限公司 Training method and device of multi-task model and storage medium
CN113298854A (en) * 2021-05-27 2021-08-24 广州柏视医疗科技有限公司 Image registration method based on mark points

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106097347A (en) * 2016-06-14 2016-11-09 福州大学 A kind of multimodal medical image registration and method for visualizing
CN107818564B (en) * 2017-10-27 2020-10-30 深圳市一图智能科技有限公司 Liver 3D medical image segmentation method
CN111640143B (en) * 2020-04-12 2023-05-30 复旦大学 Neural navigation rapid surface registration method and system based on PointNet

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104021547A (en) * 2014-05-17 2014-09-03 清华大学深圳研究生院 Three dimensional matching method for lung CT
US20170337682A1 (en) * 2016-05-18 2017-11-23 Siemens Healthcare Gmbh Method and System for Image Registration Using an Intelligent Artificial Agent
CN109064502A (en) * 2018-07-11 2018-12-21 西北工业大学 The multi-source image method for registering combined based on deep learning and artificial design features
CN111091589A (en) * 2019-11-25 2020-05-01 北京理工大学 Ultrasonic and nuclear magnetic image registration method and device based on multi-scale supervised learning
CN111931929A (en) * 2020-07-29 2020-11-13 深圳地平线机器人科技有限公司 Training method and device of multi-task model and storage medium
CN113298854A (en) * 2021-05-27 2021-08-24 广州柏视医疗科技有限公司 Image registration method based on mark points

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117474993A (en) * 2023-10-27 2024-01-30 哈尔滨工程大学 Underwater image feature point sub-pixel position estimation method and device
CN117474993B (en) * 2023-10-27 2024-05-24 哈尔滨工程大学 Underwater image feature point sub-pixel position estimation method and device

Also Published As

Publication number Publication date
CN113298854A (en) 2021-08-24
CN113298854B (en) 2022-02-01

Similar Documents

Publication Publication Date Title
WO2022247296A1 (en) Mark point-based image registration method
Mahmood et al. Unsupervised reverse domain adaptation for synthetic medical images via adversarial training
CN111476292B (en) Small sample element learning training method for medical image classification processing artificial intelligence
CN110148142B (en) Training method, device and equipment of image segmentation model and storage medium
Liu et al. Extremely dense point correspondences using a learned feature descriptor
Cao et al. Region-adaptive deformable registration of CT/MRI pelvic images via learning-based image synthesis
WO2022247218A1 (en) Image registration method based on automatic delineation
US8861891B2 (en) Hierarchical atlas-based segmentation
Banerjee et al. Fast and robust 3D ultrasound registration–block and game theoretic matching
CN113826143A (en) Feature point detection
Markova et al. Global multi-modal 2D/3D registration via local descriptors learning
Jin et al. Object recognition in medical images via anatomy-guided deep learning
Qiu et al. RSegNet: a joint learning framework for deformable registration and segmentation
WO2022188799A1 (en) Brain identifier positioning system and method
Zhang et al. Segmentation of Thoracic Organs at Risk in CT Images Combining Coarse and Fine Network.
Song et al. Combining deep learning with geometric features for image-based localization in the Gastrointestinal tract
Zhu et al. Similarity attention-based CNN for robust 3D medical image registration
CN112686932B (en) Image registration method for medical image, image processing method and medium
CN106709867A (en) Medical image registration method based on improved SURF and improved mutual information
CN116883462A (en) Medical image registration method based on LOFTR network model and improved particle swarm optimization
Kong et al. Cascade connection-based channel attention network for bidirectional medical image registration
Falta et al. Lung250M-4B: a combined 3D dataset for CT-and point cloud-based intra-patient lung registration
Fan et al. Deep feature descriptor based hierarchical dense matching for X-ray angiographic images
Jaouen et al. Regularized directional representations for medical image registration
Hernández et al. Investigating keypoint descriptors for camera relocalization in endoscopy surgery

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 22810035

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 22810035

Country of ref document: EP

Kind code of ref document: A1