CN118470222B - Medical ultrasonic image three-dimensional reconstruction method and system based on SDF diffusion - Google Patents
Medical ultrasonic image three-dimensional reconstruction method and system based on SDF diffusion Download PDFInfo
- Publication number
- CN118470222B CN118470222B CN202410917497.7A CN202410917497A CN118470222B CN 118470222 B CN118470222 B CN 118470222B CN 202410917497 A CN202410917497 A CN 202410917497A CN 118470222 B CN118470222 B CN 118470222B
- Authority
- CN
- China
- Prior art keywords
- sdf
- voxel
- ultrasonic image
- diffusion
- features
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 238000009792 diffusion process Methods 0.000 title claims abstract description 70
- 238000000034 method Methods 0.000 title claims abstract description 65
- 238000002604 ultrasonography Methods 0.000 claims abstract description 77
- 230000002441 reversible effect Effects 0.000 claims abstract description 8
- 230000000007 visual effect Effects 0.000 claims abstract description 7
- 230000008569 process Effects 0.000 claims description 32
- 230000006870 function Effects 0.000 claims description 19
- 238000013528 artificial neural network Methods 0.000 claims description 17
- 238000004422 calculation algorithm Methods 0.000 claims description 11
- 238000004364 calculation method Methods 0.000 claims description 8
- 239000011159 matrix material Substances 0.000 claims description 8
- 241000270311 Crocodylus niloticus Species 0.000 claims description 6
- 238000000605 extraction Methods 0.000 claims description 6
- 230000007704 transition Effects 0.000 claims description 6
- 230000003993 interaction Effects 0.000 claims description 5
- 238000013507 mapping Methods 0.000 claims description 5
- 239000013598 vector Substances 0.000 claims description 5
- 238000005070 sampling Methods 0.000 claims description 4
- 230000001419 dependent effect Effects 0.000 claims description 3
- 238000012549 training Methods 0.000 claims description 3
- 238000011426 transformation method Methods 0.000 claims description 3
- 238000013459 approach Methods 0.000 claims description 2
- 238000012856 packing Methods 0.000 claims 2
- 238000004321 preservation Methods 0.000 claims 1
- 238000005516 engineering process Methods 0.000 abstract description 2
- 238000012795 verification Methods 0.000 abstract description 2
- 239000000523 sample Substances 0.000 description 19
- 238000013527 convolutional neural network Methods 0.000 description 3
- 230000004927 fusion Effects 0.000 description 3
- 230000009466 transformation Effects 0.000 description 3
- 238000010586 diagram Methods 0.000 description 2
- 239000000284 extract Substances 0.000 description 2
- 238000013178 mathematical model Methods 0.000 description 2
- 238000012731 temporal analysis Methods 0.000 description 2
- 238000000700 time series analysis Methods 0.000 description 2
- 101001121408 Homo sapiens L-amino-acid oxidase Proteins 0.000 description 1
- 102100026388 L-amino-acid oxidase Human genes 0.000 description 1
- 101100012902 Saccharomyces cerevisiae (strain ATCC 204508 / S288c) FIG2 gene Proteins 0.000 description 1
- 210000003484 anatomy Anatomy 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000013079 data visualisation Methods 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 238000002474 experimental method Methods 0.000 description 1
- 238000005259 measurement Methods 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 238000012545 processing Methods 0.000 description 1
- 238000012360 testing method Methods 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T17/00—Three dimensional [3D] modelling, e.g. data description of 3D objects
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
- G06N3/084—Backpropagation, e.g. using gradient descent
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/44—Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components
- G06V10/443—Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components by matching or filtering
- G06V10/449—Biologically inspired filters, e.g. difference of Gaussians [DoG] or Gabor filters
- G06V10/451—Biologically inspired filters, e.g. difference of Gaussians [DoG] or Gabor filters with interaction between the filter responses, e.g. cortical complex cells
- G06V10/454—Integrating the filters into a hierarchical structure, e.g. convolutional neural networks [CNN]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/77—Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
- G06V10/80—Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level
- G06V10/806—Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level of extracted features
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/82—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02T—CLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
- Y02T10/00—Road transport of goods or passengers
- Y02T10/10—Internal combustion engine [ICE] based vehicles
- Y02T10/40—Engine management systems
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Evolutionary Computation (AREA)
- General Physics & Mathematics (AREA)
- Health & Medical Sciences (AREA)
- Artificial Intelligence (AREA)
- Software Systems (AREA)
- General Health & Medical Sciences (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Computing Systems (AREA)
- Biomedical Technology (AREA)
- Molecular Biology (AREA)
- Multimedia (AREA)
- Life Sciences & Earth Sciences (AREA)
- Computational Linguistics (AREA)
- Biophysics (AREA)
- Mathematical Physics (AREA)
- Data Mining & Analysis (AREA)
- Databases & Information Systems (AREA)
- Medical Informatics (AREA)
- General Engineering & Computer Science (AREA)
- Biodiversity & Conservation Biology (AREA)
- Computer Graphics (AREA)
- Geometry (AREA)
- Image Processing (AREA)
- Ultra Sonic Daignosis Equipment (AREA)
Abstract
本发明公开了一种基于SDF扩散的医学超声影像三维重建方法和系统,属于计算机视觉技术领域。首先使用在医学数据上预训练的视觉模型对输入超声图像提取特征。随后开始SDF扩散过程,反向扩散过程由完全随机噪声开始,逐步去噪以得到干净的SDF场,其中使用状态空间模型和交叉注意力对SDF特征和超声图像特征进行融合,以实现SDF场所表示三维表面与超声图像一致。经过实际验证,本发明提供的基于SDF扩散的医学超声影像三维重建方法具有高效、高精确度的特点。
The present invention discloses a method and system for three-dimensional reconstruction of medical ultrasound images based on SDF diffusion, and belongs to the field of computer vision technology. First, a visual model pre-trained on medical data is used to extract features of an input ultrasound image. Then, the SDF diffusion process is started, and the reverse diffusion process starts with completely random noise, and gradually denoises to obtain a clean SDF field, wherein the SDF features and ultrasound image features are fused using a state space model and cross attention to achieve the consistency of the SDF field representation of the three-dimensional surface with the ultrasound image. After actual verification, the medical ultrasound image three-dimensional reconstruction method based on SDF diffusion provided by the present invention has the characteristics of high efficiency and high precision.
Description
技术领域Technical Field
本发明涉及一种基于SDF扩散的医学超声影像三维重建方法和系统,属于计算机视觉技术领域。The invention relates to a method and system for three-dimensional reconstruction of medical ultrasonic images based on SDF diffusion, and belongs to the technical field of computer vision.
背景技术Background Art
超声图像三维重建是指通过获取和处理一系列二维超声图像数据,最终生成三维的解剖结构图像或体积数据,具有重要的临床意义和应用价值。物体的三维形状可以有多种的表达形式,如体素、点云、网格等,基于体素表示的方法可以使用3D卷积神经网络(CNN),它可以重建具有任意拓扑结构的对象。然而,巨大的内存需求和计算时间限制了大多数方法对高分辨率的重建结果,所以无法实现超高的精度重建。点云表示相对简单且高度灵活,但是因为点云不是规则结构,所以它不能很好地适应传统的3D CNN网络。类似的,网络可以高精度的表示三维形状,但它也是离散的,无法用传统神经网络很好的表示。Ultrasound image 3D reconstruction refers to the process of acquiring and processing a series of 2D ultrasound image data to ultimately generate 3D anatomical structure images or volume data, which has important clinical significance and application value. The 3D shape of an object can be expressed in a variety of forms, such as voxels, point clouds, grids, etc. Voxel-based representation methods can use 3D convolutional neural networks (CNNs), which can reconstruct objects with arbitrary topological structures. However, the huge memory requirements and computing time limit most methods to high-resolution reconstruction results, so ultra-high precision reconstruction cannot be achieved. Point cloud representation is relatively simple and highly flexible, but because point clouds are not regular structures, they cannot adapt well to traditional 3D CNN networks. Similarly, the network can represent 3D shapes with high precision, but it is also discrete and cannot be well represented by traditional neural networks.
扩散去噪概率模型是一类基于迭代反转马尔可夫噪声过程的生成模型。在视觉中,早期的工作将问题表述为学习变分下界,或者将其框定为优化基于分数的生成模型或连续随机过程的离散化。近来的许多工作展示出扩散模型在内容生成任务上的巨大潜力,因此将扩散用于三维重建任务也将充分发挥模型的优势。Diffusion denoising probabilistic model is a class of generative models based on iterative inversion of Markov noise process. In vision, early works formulated the problem as learning variational lower bounds, or framed it as optimizing score-based generative models or discretization of continuous random processes. Many recent works have shown the great potential of diffusion models in content generation tasks, so using diffusion for 3D reconstruction tasks will also give full play to the advantages of the model.
在超声图像三维重建领域中,传统的方法有基于位置传感器的方法,它利用连接在探头上的位置传感器(如磁位传感器、光学位置传感器等)实时跟踪探头在三维空间中的位置和姿态,并将二维超声图像在获取时的空间位置信息存储下来,然后根据这些位置信息,将多个二维图像重建为三维数据。但是这种方法需要外置位置传感器系统,大大增加了系统的复杂度和成本,也存在线缆干扰、探头遮挡等问题影响精度。基于自由手扫描的方法,利用操作者自由移动超声探头在人体表面扫描时的运动轨迹和超声波束的扫描方向等信息,估计每个二维图像的空间位置关系,从而重建三维数据。但其需要高精度的探头运动轨迹和波束方向估计,不仅计算复杂而且实时性较差。基于图像配准的方法,从一系列相邻的二维超声图像中提取灰度、梯度、特征等信息,通过相似度度量进行配准,估计图像之间的几何变换关系,最终获得三维数据。但其对图像质量、视野内灰度分布均匀性等要求较高,精度易受影响,计算复杂。In the field of three-dimensional reconstruction of ultrasound images, traditional methods include position sensor-based methods, which use position sensors (such as magnetic position sensors, optical position sensors, etc.) connected to the probe to track the position and posture of the probe in three-dimensional space in real time, and store the spatial position information of the two-dimensional ultrasound image when it is acquired. Then, based on this position information, multiple two-dimensional images are reconstructed into three-dimensional data. However, this method requires an external position sensor system, which greatly increases the complexity and cost of the system. There are also problems such as cable interference and probe occlusion that affect the accuracy. The method based on free hand scanning uses the motion trajectory of the operator's free movement of the ultrasound probe when scanning the human body surface and the scanning direction of the ultrasound beam to estimate the spatial position relationship of each two-dimensional image, thereby reconstructing three-dimensional data. However, it requires high-precision probe motion trajectory and beam direction estimation, which is not only complex in calculation but also has poor real-time performance. The method based on image registration extracts grayscale, gradient, feature and other information from a series of adjacent two-dimensional ultrasound images, performs registration through similarity measurement, estimates the geometric transformation relationship between images, and finally obtains three-dimensional data. However, it has high requirements on image quality and uniformity of grayscale distribution in the field of view, is easily affected by accuracy, and is complex in calculation.
通过对现有的超声三维重建方法进行分析总结,已有的方法存在下面几个不足:(1)需要额外的设备或复杂的算法确定探头运动轨迹和方向,增加了实施难度和成本。(2)重建过程需要连续的超声视频或一系列连续的超声图像,所需数据量大,计算复杂。(3)重建过程耗时长,且精度一般。By analyzing and summarizing the existing ultrasound 3D reconstruction methods, the existing methods have the following shortcomings: (1) Additional equipment or complex algorithms are required to determine the probe motion trajectory and direction, which increases the difficulty and cost of implementation. (2) The reconstruction process requires continuous ultrasound video or a series of continuous ultrasound images, which requires a large amount of data and complex calculations. (3) The reconstruction process is time-consuming and has average accuracy.
发明内容Summary of the invention
本发明的目的是提出一种基于SDF(Signed Distance Field符号距离场)扩散的医学超声影像三维重建方法,以提高二维超声图像三维重建的速度和准确性。The purpose of the present invention is to propose a medical ultrasound image three-dimensional reconstruction method based on SDF (Signed Distance Field) diffusion to improve the speed and accuracy of three-dimensional reconstruction of two-dimensional ultrasound images.
SDF是一种物体三维形状的隐式表示方法,SDF的本质是存储每个点到形状表面的最近距离,即将模型划出一个表面,在模型表面外侧的点数值大于0,在模型表面内侧的点数值小于0,SDF可以方便的使用3D卷积网络,也能通过行进立方体的方法转化为高精度的网格表示。SDF is an implicit representation method of the three-dimensional shape of an object. The essence of SDF is to store the shortest distance from each point to the surface of the shape, that is, to draw a surface out of the model. The point values on the outside of the model surface are greater than 0, and the point values on the inside of the model surface are less than 0. SDF can be conveniently used in 3D convolutional networks, and can also be converted into a high-precision grid representation through the marching cube method.
为实现上述发明目的,本发明采取的具体技术方案如下:In order to achieve the above-mentioned invention object, the specific technical scheme adopted by the present invention is as follows:
一种基于SDF扩散的医学超声影像三维重建方法,包括以下步骤:A method for three-dimensional reconstruction of medical ultrasound images based on SDF diffusion comprises the following steps:
S1:获取医学超声图像数据,并提取该超声图像特征,用于指导后续的扩散过程;S1: Acquire medical ultrasound image data and extract features of the ultrasound image to guide the subsequent diffusion process;
S2:随机初始化SDF场,将SDF场按照正态分布进行随机初始化即; S2: Randomly initialize the SDF field and set the SDF field Initialize randomly according to normal distribution. ;
S3:利用扩散模型进行SDF扩散过程,扩散模型包含两个关键过程:前向扩散过程和反向扩散过程;先使用状态空间模型对体素和超声图像特征进行融合,得到融合超声图像特征后的体素特征,再在体素分辨率较小的阶段使用交叉注意力对体素特征和超声图像特征进行融合;S3: The diffusion model is used to perform the SDF diffusion process. The diffusion model includes two key processes: forward diffusion process and backward diffusion process. The state space model is first used to fuse the voxel and ultrasound image features to obtain the voxel features after the fusion of ultrasound image features. Then, the cross attention is used to fuse the voxel features and ultrasound image features at a stage with a small voxel resolution.
S4:使用行进立方体提取网格,最终重建完成由超声图像指导的三角形网格模型。S4: The mesh is extracted using marching cubes and the triangular mesh model guided by the ultrasound image is finally reconstructed.
进一步的,所述S1中,首先使用大量医学超声图像数据训练视觉模型MedSAM (Medical Segment Anything Model)再利用训练好的MedSAM的图像编码器对输入单个医 学超声图像进行特征提取,并将输出映射至特征向量,在后续扩散过程中用 于指导SDF向图片特征靠近,以实现超声图像的SDF三维重建。 Furthermore, in S1, a large amount of medical ultrasound image data is first used to train the visual model MedSAM (Medical Segment Anything Model), and then the trained MedSAM image encoder is used to extract features from the input single medical ultrasound image, and the output is mapped to a feature vector , which is used to guide the SDF to approach the image features in the subsequent diffusion process to achieve SDF 3D reconstruction of ultrasound images.
进一步的,所述S2中,SDF场是一种连续的三维场景表示方式,它将一个三维物体或场景看作是一个由距离值组成的三维标量场。具体来,对于场景中的任意一点,SDF的值定义为该点到物体表面的有符号距离,即:如果该点在物体内部,SDF值为该点到最近表面的负距离值;如果该点在物体外部,SDF值为该点到最近表面的正距离值;如果该点正好在物体表面上,SDF值为0。Furthermore, in S2, the SDF field is a continuous three-dimensional scene representation method, which regards a three-dimensional object or scene as a three-dimensional scalar field composed of distance values. Specifically, for any point in the scene, the SDF value is defined as the signed distance from the point to the surface of the object, that is, if the point is inside the object, the SDF value is the negative distance value from the point to the nearest surface; if the point is outside the object, the SDF value is the positive distance value from the point to the nearest surface; if the point is exactly on the surface of the object, the SDF value is 0.
进一步的,所述S3具体包括:Furthermore, the S3 specifically includes:
S3-1:前向扩散过程(Forward Diffusion Process)是在训练时使用的过程,它将 数据逐步添加高斯噪声,直到最终被完全破坏,变成具有高斯分布的纯噪声。这个过程可以 用马尔可夫链表示,每个时间步t都会添加一定量的高斯噪声,直到最终时刻T,原始数据变 成纯噪声;给定一个数据样本,正向过程它逐渐将数据样本转换为纯高 斯噪声: S3-1: The forward diffusion process is a process used during training. It gradually adds Gaussian noise to the data until it is completely destroyed and becomes pure noise with Gaussian distribution. This process can be represented by a Markov chain. A certain amount of Gaussian noise is added at each time step t until the final time T, when the original data becomes pure noise. Given a data sample , forward process It gradually transforms the data samples into pure Gaussian noise:
(1); (1);
其中表示时间步t时的样本,表示前一个时间步的样本,为高斯 转移概率: in represents the sample at time step t, represents the sample of the previous time step, is the Gaussian transition probability:
(2); (2);
其中表示正态分布,表示噪声调度超参数,是一个可学习的系数或设置为一 个常数;正向过程允许以封闭形式在任意时间步t采样 : in represents a normal distribution, represents the noise scheduling hyperparameter, which is a learnable coefficient or set to a constant; the forward process allows sampling at any time step t in closed form :
(3); (3);
其中,;因此,直接通过下式采样: in , ;therefore, Directly sample through the following formula:
(4); (4);
其中的是一个从正态分布随机采样的噪声。Among them is a noise randomly sampled from a normal distribution.
S3-2:反向扩散过程(Reverse Diffusion Process)发生在生成时,旨在从纯噪声 样本重构原始数据,即联合分布,这个反向过程被定义为具有学习高斯跃迁的马尔 可夫链: S3-2: The reverse diffusion process occurs at generation time and aims to reconstruct the original data from pure noise samples, i.e., the joint distribution , the reverse process is defined as a Markov chain with learned Gaussian transitions:
(5); (5);
(6); (6);
其中是标准正态分布,表示由参数化的去噪函数,是一个时间步长相关的方差,设置为;从中采样,然后通过公式(6)绘 制为: in is the standard normal distribution, Indicated by Parameterized denoising function, is a time step dependent variance, set to ;from Then, we draw the for:
(7); (7);
(8); (8);
其中是一个由参数化的神经网络,它从噪声输入预测噪声。通过重复 这个过程,最终可以生成; in is a A parameterized neural network that starts with a noisy input By repeating this process, we can eventually generate ;
因此,目标函数比较了预测噪声与应用噪声之间的差异: Therefore, the objective function compares the prediction noise With applied noise The difference between:
(9); (9);
此外,通过神经网络预测而不是噪声来修改: In addition, the neural network prediction Rather than noise To modify :
(10); (10);
其中,,,是由参数化的神经网络,它预测无 噪声数据;在这种情况下,目标函数变为: in, , , Is A parameterized neural network that predicts noise-free data ; In this case, the objective function becomes:
(11); (11);
具体来说,给定一个干净的数据样本,按照正向过程中的公式(4)采样具有相 同形状的: Specifically, given a clean data sample , according to formula (4) in the forward process, we sample the same shape :
(12); (12);
然后,训练一个网络来预测无噪声数据,其 MSE 目标函数如下所示: Then, train a network To predict noise-free data, the MSE objective function is as follows:
(13); (13);
其中的即S1中提取的超声图片特征; Among them That is, the ultrasound image features extracted in S1;
在推理过程中,使用训练后的神经网络经过的过程按照公式(14)逐 步预测较小的噪声样本来生成新的SDF体素: During inference, the trained neural network is used go through The process of gradually predicting smaller noise samples according to formula (14) To generate new SDF voxels:
(14); (14);
S3-3:扩散过程使用的是一个类U形网络,每一层的体素特征分辨率减半(例如:64→32→16),通道加倍;为了使生成的SDF场表示的物体与超声图像保持一致,需要在扩散过程中对SDF特征和超声图像特征进行融合,让超声图像特征指导SDF扩散过程;在模型的顶部,由于特征分辨率较大,体素特征保持在一个较大的数量级,无法直接使用交叉注意力对体素和超声图像特征进行融合,所以使用状态空间模型对其进行处理;S3-3: The diffusion process uses a U-shaped network, where the voxel feature resolution of each layer is halved (for example: 64→32→16) and the channels are doubled; in order to make the object represented by the generated SDF field consistent with the ultrasound image, it is necessary to fuse the SDF features and the ultrasound image features during the diffusion process, so that the ultrasound image features can guide the SDF diffusion process; at the top of the model, due to the large feature resolution, the voxel features are kept at a large order of magnitude, and the cross-attention cannot be directly used to fuse the voxel and ultrasound image features, so the state space model is used to process them;
状态空间模型(State Space Model)是一种常用于时间序列分析和控制理论的数 学模型框架。它将动态系统的演化过程描述为一个状态方程和观测方程的组合。将一个一 维的序列通过一个隐藏状态映射至,这个系统使用A作为演化参数,B、C作为投 影参数: The state space model is a mathematical model framework commonly used in time series analysis and control theory. It describes the evolution of a dynamic system as a combination of a state equation and an observation equation. Through a hidden state Map to , this system uses A as the evolution parameter, and B and C as the projection parameters:
(15); (15);
(16); (16);
使用的Mamba是连续系统的离散版本,其中包括一个时间尺度参数将连续参数A、 B转换为离散参数、;常用的变换方法是零阶保持 (ZOH),定义如下: The Mamba used is a discrete version of the continuous system, which includes a time scale parameter Convert continuous parameters A and B into discrete parameters , ; The commonly used transformation method is zero-order hold (ZOH), which is defined as follows:
(17); (17);
(18); (18);
在、离散之后,公式(15)(16)可以表示为: exist , After discretization, formulas (15) and (16) can be expressed as:
(19); (19);
(20); (20);
最后,模型通过全局卷积计算输出:Finally, the model calculates the output through global convolution:
(21); (twenty one);
(22); (twenty two);
其中 M 是输入序列x的长度,是结构化卷积核。 where M is the length of the input sequence x, is a structured convolution kernel.
为了将其运用至SDF扩散过程中,将体素特征展开为一维特征,并将S3-1中的时间 步长t经过正余弦变换后线性映射至与体素特征相同大小的一维特征并与体素特征相加。 为了融合超声图像特征与体素特征,将超声图像特征经过线性映射至与体素特征相同的 通道数后与体素特征连接,得到SSM的输入x。 In order to apply it to the SDF diffusion process, the voxel features are expanded into one-dimensional features, and the time step t in S3-1 is linearly mapped to a one-dimensional feature of the same size as the voxel feature after sine and cosine transformation and added to the voxel feature. After linear mapping to the same number of channels as the voxel feature, it is connected with the voxel feature to obtain the input x of the SSM.
经过上述Mamba的计算后得到输出y,将体素特征部分取出后将一维序列重塑回体素特征的原形状,得到融合超声图像特征后的体素特征。After the above-mentioned calculation by Mamba, the output y is obtained. After the voxel feature part is taken out, the one-dimensional sequence is reshaped back to the original shape of the voxel feature to obtain the voxel feature after fusion of the ultrasound image feature.
S3-4:体素特征经过模型的顶部后,进行下采样使体素特征的分辨率下降,通道数增加,整体数据量减少,在体素分辨率较小的阶段使用交叉注意力对体素特征和超声图像特征进行融合。S3-4: After the voxel feature passes through the top of the model, it is downsampled to reduce the resolution of the voxel feature, increase the number of channels, and reduce the overall data volume. At the stage with smaller voxel resolution, cross attention is used to fuse the voxel feature and ultrasound image feature.
注意力函数描述为将查询和一组键值对映射到输出,其中查询、键、值和输出都是向量,输出为值的加权和。在实践中,同时计算一组查询的注意力函数,打包成一个矩阵Q,键和值也打包到矩阵K和V中;计算输出矩为:The attention function is described as mapping a query and a set of key-value pairs to an output, where the query, key, value, and output are all vectors, and the output is the weighted sum of the values. In practice, the attention function of a set of queries is calculated simultaneously, packaged into a matrix Q, and the keys and values are also packaged into matrices K and V; the output matrix is calculated as:
(23); (twenty three);
多头注意力允许模型共同关注不同位置的不同表示子空间的信息;其公式如下:Multi-head attention allows the model to jointly focus on information in different representation subspaces at different positions; its formula is as follows:
(24); (twenty four);
(25); (25);
其中concat表示拼接操作,均是可学习的权重矩阵。 Where concat represents the concatenation operation. are all learnable weight matrices.
对于网格中的体素,将其中心投影到图像上,得到投影坐标;选择靠近的相邻 图像块与交互,因为它们的特征极有可能影响由控制的局部几何形状;邻域图像补丁集 用表示,选择如下:如果与中心之间的距离小于距离阈值,则补丁属于。 For voxels in the grid , project its center onto the image and get the projection coordinates ; Select Close The adjacent image blocks are interactions, because their characteristics are likely to affect Controlled local geometry; neighborhood image patch sets are used Indicates that the selection is as follows: If and The distance between centers is less than the distance threshold , then the patch belong .
使用多头自注意力对处的体素特征和属于中的补丁的超声图像补丁特征 集之间的特征交互进行建模,如下所示: Use multi-head self-attention Voxel features at and belongs to Ultrasound image patch feature set of patches in The feature interactions between are modeled as follows:
(26); (26);
(27); (27);
这里,是标准的多头注意力操作,、、是可学习的矩阵,M 是视图投影引起的注意力计算的掩码,对体素和超声图像块使用绝对位置编码。 here, It is a standard multi-head attention operation. , , is a learnable matrix, M is the mask for attention computation induced by view projection, using absolute position encoding for voxels and ultrasound image patches.
经过交叉注意力后进一步将超声图像特征与体素特征进行了融合,此时输出的体素特征就是高精度的SDF场。After cross attention, the ultrasound image features are further fused with the voxel features. At this time, the output voxel features are high-precision SDF fields.
进一步的,S4中的行进立方体算法是一种用于从三维标量场中提取等值面(等值曲面)的经典算法,广泛应用于体数据可视化、几何重建等领域。算法步骤:设定一个等值面阈值T,遍历SDF场中所有立方体单元C:Furthermore, the marching cube algorithm in S4 is a classic algorithm for extracting isosurfaces (isosurfaces) from three-dimensional scalar fields, and is widely used in volume data visualization, geometric reconstruction and other fields. Algorithm steps: Set an isosurface threshold T and traverse all cube cells C in the SDF field:
(1) 根据C的8个顶点标量值与T的大小关系计算顶点状态码index;(1) Calculate the vertex state code index based on the relationship between the 8 vertex scalar values of C and T;
(2) 根据index在预定义的情况表中查找对应的交点配置pattern;(2) Search the corresponding intersection configuration pattern in the predefined situation table according to the index;
(3) 利用线性插值在立方体的12条边上计算与等值面的交点;(3) Use linear interpolation to calculate the intersection points with the isosurfaces on the 12 edges of the cube;
(4) 根据pattern连接相应交点形成一个或多个三角形;(4) Connect the corresponding intersection points according to the pattern to form one or more triangles;
(5) 将三角形加入网格表面数据。(5) Add triangles to the mesh surface data.
一种基于SDF扩散的医学超声影像三维重建系统,由超声图像特征提取模块、SDF扩散模块、网格生成模块组成;A medical ultrasound image 3D reconstruction system based on SDF diffusion, which consists of an ultrasound image feature extraction module, an SDF diffusion module, and a grid generation module;
所述超声图像特征提取模块:该模块用于提取输入超声图片的特征,使用经过大量医学图像数据预训练的视觉模型SAM-Med2D的图像编码器对输入超声图像进行特征提取,为后续扩散过程提供丰富准确的特征指导。The ultrasound image feature extraction module: This module is used to extract the features of the input ultrasound picture, and uses the image encoder of the visual model SAM-Med2D that has been pre-trained with a large amount of medical image data to extract features from the input ultrasound image, providing rich and accurate feature guidance for the subsequent diffusion process.
所述SDF扩散模块:该模块由一个嵌入状态空间模型和交叉注意力的U形神经网络组成,用于执行SDF扩散过程,从随机噪声开始逐步去噪以得到干净的SDF场,扩散过程中由状态空间模型和交叉注意力对体素特征和超声图片特征进行融合,实现重建网格与超声图片特征保持一致的目的。The SDF diffusion module: This module consists of a U-shaped neural network embedded with a state space model and cross attention, which is used to perform the SDF diffusion process, starting from random noise and gradually denoising to obtain a clean SDF field. During the diffusion process, the state space model and cross attention are used to fuse the voxel features and the ultrasound image features to achieve the purpose of keeping the reconstructed grid consistent with the ultrasound image features.
所述网格生成模块:该模块执行行进立方体算法,从重建完成的SDF场中提取高精度的三维网格。The mesh generation module: This module executes the marching cubes algorithm to extract a high-precision three-dimensional mesh from the reconstructed SDF field.
本发明的优点和有益效果:Advantages and beneficial effects of the present invention:
本发明将状态空间模型和交叉注意力引入至SDF扩散过程中,将超声图像特征与SDF特征充分融合,实现了一个端到端的快速且准确的超声图像三维重建。在一定程度上解决了超声图像三维重建需要额外的设备或复杂的算法确定探头运动轨迹和方向、耗时长等问题。The present invention introduces the state space model and cross attention into the SDF diffusion process, fully integrates the ultrasound image features with the SDF features, and realizes an end-to-end fast and accurate ultrasound image 3D reconstruction. To a certain extent, it solves the problems that ultrasound image 3D reconstruction requires additional equipment or complex algorithms to determine the probe motion trajectory and direction, and takes a long time.
经过实际验证,本发明提供的超声图像三维重建方法具有高效、高精确度的特点。Practical verification has shown that the ultrasonic image three-dimensional reconstruction method provided by the present invention has the characteristics of high efficiency and high precision.
附图说明BRIEF DESCRIPTION OF THE DRAWINGS
图1是本发明的整体流程图。FIG. 1 is an overall flow chart of the present invention.
图2是本发明的框架图。FIG. 2 is a framework diagram of the present invention.
图3是本发明的结果图。FIG. 3 is a diagram showing the results of the present invention.
具体实施方式DETAILED DESCRIPTION
下面将结合附图1-3及与具体实施例对本发明做进一步说明。The present invention will be further described below in conjunction with Figures 1-3 and specific embodiments.
实施例1:Embodiment 1:
一种基于SDF扩散的医学超声影像三维重建方法,整体流程如图1所示,包括如下步骤:A method for 3D reconstruction of medical ultrasound images based on SDF diffusion, the overall process is shown in FIG1 , and includes the following steps:
S1:使用经过大量医学图像数据预训练的视觉模型SAM-Med2D的图像编码器对输 入单视图进行特征提取,输出特征向量,在后续扩散过程中在状态空间模型 和交叉注意力中与SDF特征进行融合,实现指导三维重建的目的,以实现超声图像三维重 建。 S1: Use the image encoder of SAM-Med2D, a visual model pre-trained with a large amount of medical image data, to extract features from the input single view and output a feature vector , in the subsequent diffusion process, it is fused with the SDF features in the state space model and cross attention to achieve the purpose of guiding 3D reconstruction, so as to realize 3D reconstruction of ultrasound images.
S2:随机初始化体素SDF场,其中的SDF全称是Signed Distance Function(带符号 距离函数)。SDF场是一种连续的三维场景表示方式,它将一个三维物体或场景看作是一个 由距离值组成的三维标量场。具体来说,对于场景中的任意一点,SDF的值定义为该点到物 体表面的有符号距离,即:如果该点在物体内部,SDF值为该点到最近表面的负距离值;如果 该点在物体外部,SDF值为该点到最近表面的正距离值;如果该点正好在物体表面上,SDF值 为0。初始化时,将SDF场按照正态分布进行随机初始化即。 S2: Randomly initialize the voxel SDF field, where SDF stands for Signed Distance Function. The SDF field is a continuous three-dimensional scene representation that treats a three-dimensional object or scene as a three-dimensional scalar field composed of distance values. Specifically, for any point in the scene, the value of SDF is defined as the signed distance from the point to the surface of the object, that is: if the point is inside the object, the SDF value is the negative distance from the point to the nearest surface; if the point is outside the object, the SDF value is the positive distance from the point to the nearest surface; if the point is exactly on the surface of the object, the SDF value is 0. During initialization, the SDF field is initialized. Initialize randomly according to normal distribution. .
S3:对SDF场进行扩散过程。扩散模型(Diffusion Models)是一种新兴的深度生成模型,它利用可逆的扩散过程(Diffusion Process)训练神经网络,从噪声分布中生成所需的数据。生成式扩散模型包含两个关键过程:前向扩散过程和反向扩散过程。具体包括:S3: Diffusion process is performed on the SDF field. Diffusion Models is an emerging deep generative model that uses a reversible diffusion process to train a neural network to generate the required data from the noise distribution. The generative diffusion model contains two key processes: forward diffusion process and reverse diffusion process. Specifically, it includes:
S3-1:前向扩散过程(Forward Diffusion Process)是在训练时使用的过程,它将 数据逐步添加高斯噪声,直到最终被完全破坏,变成具有高斯分布的纯噪声。这个过程可以 用马尔可夫链表示,每个时间步t都会添加一定量的高斯噪声,直到最终时刻T,原始数据变 成纯噪声;给定一个数据样本,正向过程它逐渐将数据样本转换为纯高 斯噪声: S3-1: The forward diffusion process is a process used during training. It gradually adds Gaussian noise to the data until it is completely destroyed and becomes pure noise with Gaussian distribution. This process can be represented by a Markov chain. A certain amount of Gaussian noise is added at each time step t until the final time T, when the original data becomes pure noise. Given a data sample , forward process It gradually transforms the data samples into pure Gaussian noise:
(1); (1);
其中表示时间步t时的样本,表示前一个时间步的样本,为高斯 转移概率: in represents the sample at time step t, represents the sample of the previous time step, is the Gaussian transition probability:
(2); (2);
其中表示正态分布,表示噪声调度超参数,是一个可学习的系数或设置为一 个常数;正向过程允许以封闭形式在任意时间步t采样 : in represents a normal distribution, represents the noise scheduling hyperparameter, which is a learnable coefficient or set to a constant; the forward process allows sampling at any time step t in closed form :
(3); (3);
其中,;因此,直接通过下式采样: in , ;therefore, Directly sample through the following formula:
(4); (4);
其中的是一个从正态分布随机采样的噪声。 Among them is a noise randomly sampled from a normal distribution.
S3-2:反向扩散过程(Reverse Diffusion Process)发生在生成时,旨在从纯噪声 样本重构原始数据,即联合分布,这个反向过程被定义为具有学习高斯跃迁的马尔 可夫链: S3-2: The reverse diffusion process occurs at generation time and aims to reconstruct the original data from pure noise samples, i.e., the joint distribution , the reverse process is defined as a Markov chain with learned Gaussian transitions:
(5); (5);
(6); (6);
其中是标准正态分布,表示由参数化的去噪函数,是一个时间步长相关的方差,设置为;从中采样,然后通过公式(6)绘 制为: in is the standard normal distribution, Indicated by Parameterized denoising function, is a time step dependent variance, set to ;from Then, we draw the for:
(7); (7);
(8); (8);
其中是一个由参数化的神经网络,它从噪声输入预测噪声。通过重复 这个过程,最终可以生成; in is a A parameterized neural network that starts with a noisy input By repeating this process, we can eventually generate ;
因此,目标函数比较了预测噪声与应用噪声之间的差异: Therefore, the objective function compares the prediction noise With applied noise The difference between:
(9); (9);
此外,通过神经网络预测而不是噪声来修改: In addition, the neural network prediction Rather than noise To modify :
(10); (10);
其中,,,是由参数化的神经网络,它预测无 噪声数据;在这种情况下,目标函数变为: in, , , Is A parameterized neural network that predicts noise-free data ; In this case, the objective function becomes:
(11); (11);
具体来说,给定一个干净的数据样本,按照正向过程中的公式(4)采样具有相 同形状的: Specifically, given a clean data sample , according to formula (4) in the forward process, we sample the same shape :
(12); (12);
然后,训练一个网络来预测无噪声数据,其 MSE 目标函数如下所示: Then, train a network To predict noise-free data, the MSE objective function is as follows:
(13); (13);
其中的即S1中提取的超声图片特征; Among them That is, the ultrasound image features extracted in S1;
在推理过程中,使用训练后的神经网络经过的过程按照公式(14)逐 步预测较小的噪声样本来生成新的SDF体素: During inference, the trained neural network is used go through The process of gradually predicting smaller noise samples according to formula (14) To generate new SDF voxels:
(14); (14);
S3-3:扩散过程使用的是一个类U形网络,每一层的体素特征分辨率减半(例如:64→32→16),通道加倍;为了使生成的SDF场表示的物体与超声图像保持一致,需要在扩散过程中对SDF特征和超声图像特征进行融合,让超声图像特征指导SDF扩散过程;在模型的顶部,由于特征分辨率较大,体素特征保持在一个较大的数量级,无法直接使用交叉注意力对体素和超声图像特征进行融合,所以使用状态空间模型对其进行处理;S3-3: The diffusion process uses a U-shaped network, where the voxel feature resolution of each layer is halved (for example: 64→32→16) and the channels are doubled; in order to make the object represented by the generated SDF field consistent with the ultrasound image, it is necessary to fuse the SDF features and the ultrasound image features during the diffusion process, so that the ultrasound image features can guide the SDF diffusion process; at the top of the model, due to the large feature resolution, the voxel features are kept at a large order of magnitude, and the cross-attention cannot be directly used to fuse the voxel and ultrasound image features, so the state space model is used to process them;
状态空间模型(State Space Model)是一种常用于时间序列分析和控制理论的数 学模型框架。它将动态系统的演化过程描述为一个状态方程和观测方程的组合。将一个一 维的序列通过一个隐藏状态映射至,这个系统使用A作为演化参数,B、C作为投 影参数: The state space model is a mathematical model framework commonly used in time series analysis and control theory. It describes the evolution of a dynamic system as a combination of a state equation and an observation equation. Through a hidden state Map to , this system uses A as the evolution parameter, and B and C as the projection parameters:
(15); (15);
(16); (16);
使用的Mamba是连续系统的离散版本,其中包括一个时间尺度参数将连续参数A、 B转换为离散参数、;常用的变换方法是零阶保持 (ZOH),定义如下: The Mamba used is a discrete version of the continuous system, which includes a time scale parameter Convert continuous parameters A and B into discrete parameters , ; The commonly used transformation method is zero-order hold (ZOH), which is defined as follows:
(17); (17);
(18); (18);
在、离散之后,公式(15)(16)可以表示为: exist , After discretization, formulas (15) and (16) can be expressed as:
(19); (19);
(20); (20);
最后,模型通过全局卷积计算输出:Finally, the model calculates the output through global convolution:
(21); (twenty one);
(22 (twenty two
其中 M 是输入序列x的长度,是结构化卷积核。 where M is the length of the input sequence x, is a structured convolution kernel.
为了将其运用至SDF扩散过程中,将体素特征展开为一维特征,并将S3-1中的时间 步长t经过正余弦变换后线性映射至与体素特征相同大小的一维特征并与体素特征相加。 为了融合超声图像特征与体素特征,将超声图像特征经过线性映射至与体素特征相同的 通道数后与体素特征连接,得到SSM的输入x。 In order to apply it to the SDF diffusion process, the voxel features are expanded into one-dimensional features, and the time step t in S3-1 is linearly mapped to a one-dimensional feature of the same size as the voxel feature after sine and cosine transformation and added to the voxel feature. After linear mapping to the same number of channels as the voxel feature, it is connected with the voxel feature to obtain the input x of the SSM.
经过上述Mamba的计算后得到输出y,将体素特征部分取出后将一维序列重塑回体素特征的原形状,得到融合超声图像特征后的体素特征。After the above-mentioned calculation by Mamba, the output y is obtained. After the voxel feature part is taken out, the one-dimensional sequence is reshaped back to the original shape of the voxel feature to obtain the voxel feature after fusion of the ultrasound image feature.
S3-4:体素特征经过模型的顶部后,进行下采样使体素特征的分辨率下降,通道数增加,整体数据量减少,在体素分辨率较小的阶段使用交叉注意力对体素特征和超声图像特征进行融合。S3-4: After the voxel feature passes through the top of the model, it is downsampled to reduce the resolution of the voxel feature, increase the number of channels, and reduce the overall data volume. At the stage with smaller voxel resolution, cross attention is used to fuse the voxel feature and ultrasound image feature.
注意力函数描述为将查询和一组键值对映射到输出,其中查询、键、值和输出都是向量,输出为值的加权和。在实践中,同时计算一组查询的注意力函数,打包成一个矩阵Q,键和值也打包到矩阵K和V中;计算输出矩为:The attention function is described as mapping a query and a set of key-value pairs to an output, where the query, key, value, and output are all vectors, and the output is the weighted sum of the values. In practice, the attention function of a set of queries is calculated simultaneously, packaged into a matrix Q, and the keys and values are also packaged into matrices K and V; the output matrix is calculated as:
(23); (twenty three);
多头注意力允许模型共同关注不同位置的不同表示子空间的信息;其公式如下:Multi-head attention allows the model to jointly focus on information in different representation subspaces at different positions; its formula is as follows:
(24); (twenty four);
(25); (25);
其中concat表示拼接操作,均是可学习的权重矩阵。 Where concat represents the concatenation operation. are all learnable weight matrices.
对于网格中的体素,将其中心投影到图像上,得到投影坐标;选择靠近的相邻 图像块与交互,因为它们的特征极有可能影响由控制的局部几何形状;邻域图像补丁集 用表示,选择如下:如果与中心之间的距离小于距离阈值,则补丁属于。For voxels in the grid , project its center onto the image and get the projection coordinates ; Select Close The adjacent image blocks are interactions, because their characteristics are likely to affect Controlled local geometry; neighborhood image patch sets are used Indicates that the selection is as follows: If and The distance between centers is less than the distance threshold , then the patch belong .
使用多头自注意力对处的体素特征和属于中的补丁的超声图像补丁特征 集之间的特征交互进行建模,如下所示: Use multi-head self-attention Voxel features at and belongs to Ultrasound image patch feature set of patches in The feature interactions between are modeled as follows:
(26); (26);
(27); (27);
这里,是标准的多头注意力操作,、、是可学习的矩阵,M 是视图投影引起的注意力计算的掩码,对体素和超声图像块使用绝对位置编码。 here, It is a standard multi-head attention operation. , , is a learnable matrix, M is the mask for attention computation induced by view projection, using absolute position encoding for voxels and ultrasound image patches.
经过交叉注意力后进一步将超声图像特征与体素特征进行了融合,此时输出的体素特征就是高精度的SDF场。After cross attention, the ultrasound image features are further fused with the voxel features. At this time, the output voxel features are high-precision SDF fields.
S4:执行行进立方体算法从SDF场中提取网格。算法步骤:设定一个等值面阈值T,遍历SDF场中所有立方体单元C:S4: Execute the marching cube algorithm to extract the grid from the SDF field. Algorithm steps: Set an isosurface threshold T and traverse all cube cells C in the SDF field:
(1) 根据C的8个顶点标量值与T的大小关系计算顶点状态码index;(1) Calculate the vertex state code index based on the relationship between the 8 vertex scalar values of C and T;
(2) 根据index在预定义的情况表中查找对应的交点配置pattern;(2) Search the corresponding intersection configuration pattern in the predefined situation table according to the index;
(3) 利用线性插值在立方体的12条边上计算与等值面的交点;(3) Use linear interpolation to calculate the intersection points with the isosurfaces on the 12 edges of the cube;
(4) 根据pattern连接相应交点形成一个或多个三角形;(4) Connect the corresponding intersection points according to the pattern to form one or more triangles;
(5) 将三角形加入网格表面数据;(5) Add triangles to the mesh surface data;
得到最终重建完成的高精度由超声图像指导的三角形网格模型。The final reconstructed high-precision triangular mesh model guided by the ultrasound image is obtained.
实施例2:该实施例以实施例1为基本方法,进行模块设计。Embodiment 2: This embodiment uses embodiment 1 as the basic method to carry out module design.
一种基于SDF扩散的医学超声影像三维重建系统,由超声图像特征提取模块、SDF扩散模块、网格生成模块组成,如图2所示,以下对各部分进行详细说明:A medical ultrasound image 3D reconstruction system based on SDF diffusion is composed of an ultrasound image feature extraction module, an SDF diffusion module, and a grid generation module, as shown in FIG2 . The following describes each part in detail:
超声图像特征提取模块:该模块用于提取输入超声图片的特征,使用经过大量医学图像数据预训练的视觉模型SAM-Med2D的图像编码器对输入超声图像进行特征提取,为后续扩散过程提供丰富准确的特征指导。Ultrasound image feature extraction module: This module is used to extract the features of the input ultrasound image. It uses the image encoder of the visual model SAM-Med2D, which has been pre-trained with a large amount of medical image data, to extract features from the input ultrasound image, providing rich and accurate feature guidance for the subsequent diffusion process.
SDF扩散模块:该模块由一个嵌入状态空间模型和交叉注意力的U形神经网络组成,用于执行SDF扩散过程,从随机噪声开始逐步去噪以得到干净的SDF场,扩散过程中由状态空间模型和交叉注意力对体素特征和超声图片特征进行融合,实现重建网格与超声图片特征保持一致的目的。SDF diffusion module: This module consists of a U-shaped neural network embedded with a state-space model and cross-attention, which is used to perform the SDF diffusion process, starting from random noise and gradually denoising to obtain a clean SDF field. During the diffusion process, the state-space model and cross-attention are used to fuse the voxel features and ultrasound image features to achieve the purpose of keeping the reconstructed grid consistent with the ultrasound image features.
网格生成模块:该模块执行行进立方体算法,从重建完成的SDF场中提取高精度的三维网格。Mesh Generation Module: This module executes the marching cubes algorithm to extract a high-precision 3D mesh from the reconstructed SDF field.
实施例3:该实施例以上述方法和系统为基础进行实例验证:Embodiment 3: This embodiment is verified by example based on the above method and system:
为了验证本发明超声图像三维重建的准确度与效率,在USOVA3D数据集上进行了实验。In order to verify the accuracy and efficiency of the three-dimensional reconstruction of ultrasound images of the present invention, experiments were conducted on the USOVA3D dataset.
该实施例以上述方法和系统为基础进行实例验证,为了验证本发明提出的超声图像SDF三维重建的准确性,在USOVA3D数据集上进行测试。得到如图3所示的结果,本发明方法(Unet-mamba)能较好的根据输入超声图像生成三维形状,且输出用时较短。本发明构建的模型在超声图像的SDF三维重建的精度较高且速度较快。This embodiment is verified by examples based on the above method and system. In order to verify the accuracy of the SDF three-dimensional reconstruction of ultrasound images proposed by the present invention, the test is conducted on the USOVA3D data set. The result shown in Figure 3 is obtained. The method of the present invention (Unet-mamba) can generate three-dimensional shapes based on the input ultrasound image well, and the output time is short. The model constructed by the present invention has high accuracy and fast speed in the SDF three-dimensional reconstruction of ultrasound images.
以上计划方案,仅为本发明中的实施方法,但本发明的保护范围不限于此,所有熟悉该技术的人员在本发明所披露的技术范围以内,可理解想到的替换或者变换,都应该包含在本发明的保护范围之内,所以,本发明的保护范围应以权利要求书的保护范围为准。The above plans are only implementation methods of the present invention, but the protection scope of the present invention is not limited thereto. All replacements or changes that can be understood and thought of by persons familiar with the technology within the technical scope disclosed in the present invention should be included in the protection scope of the present invention. Therefore, the protection scope of the present invention shall be based on the protection scope of the claims.
Claims (4)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202410917497.7A CN118470222B (en) | 2024-07-10 | 2024-07-10 | Medical ultrasonic image three-dimensional reconstruction method and system based on SDF diffusion |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202410917497.7A CN118470222B (en) | 2024-07-10 | 2024-07-10 | Medical ultrasonic image three-dimensional reconstruction method and system based on SDF diffusion |
Publications (2)
Publication Number | Publication Date |
---|---|
CN118470222A CN118470222A (en) | 2024-08-09 |
CN118470222B true CN118470222B (en) | 2024-09-06 |
Family
ID=92162282
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202410917497.7A Active CN118470222B (en) | 2024-07-10 | 2024-07-10 | Medical ultrasonic image three-dimensional reconstruction method and system based on SDF diffusion |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN118470222B (en) |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN116580172A (en) * | 2023-05-19 | 2023-08-11 | 湖南大学 | Diffusion model-based three-dimensional shape generation method applied to industrial vision requirements |
CN116912419A (en) * | 2023-07-21 | 2023-10-20 | 广州大学 | Three-dimensional human body reconstruction method, system, equipment and medium based on diffusion model |
Family Cites Families (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11087459B2 (en) * | 2015-08-14 | 2021-08-10 | Elucid Bioimaging Inc. | Quantitative imaging for fractional flow reserve (FFR) |
CN117689805A (en) * | 2023-05-09 | 2024-03-12 | 北京航空航天大学 | Large-scale cloud scene simulation method based on noise and particles |
CN117115355A (en) * | 2023-09-07 | 2023-11-24 | 上海微创电生理医疗科技股份有限公司 | Three-dimensional ultrasonic modeling method, system, electronic device and readable storage medium |
CN117541471B (en) * | 2023-11-09 | 2024-06-07 | 西安电子科技大学 | SPH heuristic PG-SPECT image super-resolution reconstruction method |
CN118247414A (en) * | 2024-03-26 | 2024-06-25 | 杭州电子科技大学 | Small sample image reconstruction method based on combined diffusion texture constraint nerve radiation field |
CN117953180B (en) * | 2024-03-26 | 2024-10-08 | 厦门大学 | A text-to-3D object generation method based on bimodal latent variable diffusion |
CN118196121B (en) * | 2024-04-08 | 2024-09-20 | 兰州交通大学 | Breast ultrasound image segmentation method based on denoising diffusion probability model |
CN118229886A (en) * | 2024-04-28 | 2024-06-21 | 中国人民解放军空军工程大学 | In-situ three-dimensional reconstruction method, device and storage medium for engine blades based on deep learning and neural radiation field |
-
2024
- 2024-07-10 CN CN202410917497.7A patent/CN118470222B/en active Active
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN116580172A (en) * | 2023-05-19 | 2023-08-11 | 湖南大学 | Diffusion model-based three-dimensional shape generation method applied to industrial vision requirements |
CN116912419A (en) * | 2023-07-21 | 2023-10-20 | 广州大学 | Three-dimensional human body reconstruction method, system, equipment and medium based on diffusion model |
Also Published As
Publication number | Publication date |
---|---|
CN118470222A (en) | 2024-08-09 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
KR102662620B1 (en) | 3D object reconstruction | |
CN110288695B (en) | Surface reconstruction method of single-frame image 3D model based on deep learning | |
CN102855470B (en) | Estimation method of human posture based on depth image | |
Venkatesh et al. | Deep implicit surface point prediction networks | |
CN101907459B (en) | Monocular video based real-time posture estimation and distance measurement method for three-dimensional rigid body object | |
Paulsen et al. | Markov random field surface reconstruction | |
CN110188598A (en) | A Real-time Hand Pose Estimation Method Based on MobileNet-v2 | |
JP2021535499A (en) | Face Posture Estimate / 3D Face Reconstruction Methods, Devices, and Electronic Devices | |
Demir et al. | Skelneton 2019: Dataset and challenge on deep learning for geometric shape understanding | |
CN113781640A (en) | A method for establishing a 3D face reconstruction model based on weakly supervised learning and its application | |
CN113052955A (en) | Point cloud completion method, system and application | |
Pons-Moll et al. | Model-based pose estimation | |
CN112184912A (en) | Multi-metric three-dimensional face reconstruction method based on parameterized model and position map | |
Goldenstein et al. | Statistical cue integration in dag deformable models | |
CN113888697A (en) | A three-dimensional reconstruction method in the state of two-hand interaction | |
CN116051696A (en) | Reconstruction method and device of human body implicit model capable of being re-illuminated | |
Zhu et al. | Nonlocal low-rank point cloud denoising for 3-D measurement surfaces | |
Han et al. | A double branch next-best-view network and novel robot system for active object reconstruction | |
Wang et al. | Arbitrary spatial trajectory reconstruction based on a single inertial sensor | |
Wu et al. | An unsupervised real-time framework of human pose tracking from range image sequences | |
CN116597112A (en) | 3D reconstruction method and device based on geometric prior network | |
CN118470222B (en) | Medical ultrasonic image three-dimensional reconstruction method and system based on SDF diffusion | |
US20240212279A1 (en) | Detailed 3d object reconstruction method and apparatus using laplacian coordinates | |
Chen et al. | Learning shape priors for single view reconstruction | |
CN118071932A (en) | Three-dimensional static scene image reconstruction method and system |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |