UNet

UNet
  • 文章类型: Journal Article
    腹部有多个重要器官,与各种疾病有关,对人类健康构成重大风险。及早发现腹部器官状况,可以及时进行干预和治疗,防止患者健康恶化。分割腹部器官有助于医生更准确地诊断器官病变。然而,腹部器官的解剖结构相对复杂,器官相互重叠,共享类似的功能,从而为细分任务提出了挑战。在真实的医疗场景中,模型必须展示实时和低延迟功能,需要在最小化参数数量的同时提高分割精度。研究人员开发了各种腹部器官分割方法,从卷积神经网络(CNN)到变形金刚。然而,这些方法在准确识别器官分割边界时经常遇到困难。MetaFormer抽象了变形金刚的框架,不包括多头自我关注,为解决计算机视觉问题和克服视觉变形金刚和CNN骨干网络的局限性提供了新的视角。为了进一步提高分割效果,我们提出了一个U形网络,集成SEFormer和深度级联上采样(dCUP)作为编码器和解码器,分别,进入UNet结构,名为SEF-UNet。SEFormer将挤压和激励模块与深度可分离卷积相结合,实例化MetaFormer框架,增强局部细节和纹理信息的捕获,从而提高边缘分割精度。dCUP在上采样过程中进一步集成了浅层和深层信息层。我们的模型显着提高了分割精度,同时减少了参数计数,并在分割彼此重叠的器官边缘方面表现出卓越的性能,从而在真实的医疗场景中提供潜在的部署。
    The abdomen houses multiple vital organs, which are associated with various diseases posing significant risks to human health. Early detection of abdominal organ conditions allows for timely intervention and treatment, preventing deterioration of patients\' health. Segmenting abdominal organs aids physicians in more accurately diagnosing organ lesions. However, the anatomical structures of abdominal organs are relatively complex, with organs overlapping each other, sharing similar features, thereby presenting challenges for segmentation tasks. In real medical scenarios, models must demonstrate real-time and low-latency features, necessitating an improvement in segmentation accuracy while minimizing the number of parameters. Researchers have developed various methods for abdominal organ segmentation, ranging from convolutional neural networks (CNNs) to Transformers. However, these methods often encounter difficulties in accurately identifying organ segmentation boundaries. MetaFormer abstracts the framework of Transformers, excluding the multi-head Self-Attention, offering a new perspective for solving computer vision problems and overcoming the limitations of Vision Transformers and CNN backbone networks. To further enhance segmentation effectiveness, we propose a U-shaped network, integrating SEFormer and depthwise cascaded upsampling (dCUP) as the encoder and decoder, respectively, into the UNet structure, named SEF-UNet. SEFormer combines Squeeze-and-Excitation modules with depthwise separable convolutions, instantiating the MetaFormer framework, enhancing the capture of local details and texture information, thereby improving edge segmentation accuracy. dCUP further integrates shallow and deep information layers during the upsampling process. Our model significantly improves segmentation accuracy while reducing the parameter count and exhibits superior performance in segmenting organ edges that overlap each other, thereby offering potential deployment in real medical scenarios.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    最近,基于编码器-解码器架构的ViT和CNN已经成为医学图像分割领域的主导模型。然而,它们中的每一个都有一些不足:(1)考虑到较长距离,CNN很难捕获两个位置之间的相互作用。(2)ViT无法获取局部上下文信息的交互,计算复杂度高。为了优化上述不足,我们提出了一种新的医学图像分割网络,称为FCSU-Net。FCSU-Net使用提出的多尺度特征块的协作融合,使网络获得更丰富,更准确的特征。此外,FCSU-Net通过FFF(全尺度特征融合)结构融合全尺度特征信息,而不是简单的跳过连接,并通过CS(Cross-dimensionSelf-attention)机制在多个维度上建立长程依赖关系。同时,每个维度都是相辅相成的。此外,CS机制具有卷积捕获局部上下文权重的优点。最后,FCSU-Net在多个数据集上进行了验证,结果表明,FCSU-Net不仅参数数量相对较少,而且还具有领先的细分性能。
    Recently, ViT and CNNs based on encoder-decoder architecture have become the dominant model in the field of medical image segmentation. However, there are some deficiencies for each of them: (1) It is difficult for CNNs to capture the interaction between two locations with consideration of the longer distance. (2) ViT cannot acquire the interaction of local context information and carries high computational complexity. To optimize the above deficiencies, we propose a new network for medical image segmentation, which is called FCSU-Net. FCSU-Net uses the proposed collaborative fusion of multi-scale feature block that enables the network to obtain more abundant and more accurate features. In addition, FCSU-Net fuses full-scale feature information through the FFF (Full-scale Feature Fusion) structure instead of simple skip connections, and establishes long-range dependencies on multiple dimensions through the CS (Cross-dimension Self-attention) mechanism. Meantime, every dimension is complementary to each other. Also, CS mechanism has the advantage of convolutions capturing local contextual weights. Finally, FCSU-Net is validated on several datasets, and the results show that FCSU-Net not only has a relatively small number of parameters, but also has a leading segmentation performance.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    对抗性训练在增强图像的视觉真实感方面备受关注,但其在临床影像学中的疗效尚未被探讨。这项工作调查了临床背景下的对抗训练,通过在OASIS-1数据集上训练206网络来改善低分辨率和低信噪比(SNR)磁共振图像。每个网络都对应于感知和对抗性损失权重以及不同的学习率值的不同组合。对于每个感知损失加权,我们确定了其相应的对抗性损失权重,将结构差异降至最低。每个最佳加权对抗损失产生平均1.5%的SSIM减少。我们进一步引入了一组新的度量标准来评估其他临床相关的图像特征:梯度误差(GE),用于测量结构差异;清晰度计算边缘清晰度;边缘对比度误差(ECE),用于量化边缘周围像素分布的任何失真。包括对抗性损失增加了视觉检查中的结构增强,这与统计学上一致的GE减少相关(p值<<0.05)。这也导致了锐度的增加;然而,统计学显著性水平取决于感知损失权重.此外,对抗性损失导致ECE减少,感知损失权重较小,当这些权重较高时,显示出不显著的增加(p值>>0.05),证明增加的清晰度不会不利地扭曲图像边缘周围的像素分布。这些研究清楚地表明,对抗训练显着提高了MRI增强管道的性能,并强调需要对超参数优化进行系统研究,并研究替代图像质量度量。
    Adversarial training has attracted much attention in enhancing the visual realism of images, but its efficacy in clinical imaging has not yet been explored. This work investigated adversarial training in a clinical context, by training 206 networks on the OASIS-1 dataset for improving low-resolution and low signal-to-noise ratio (SNR) magnetic resonance images. Each network corresponded to a different combination of perceptual and adversarial loss weights and distinct learning rate values. For each perceptual loss weighting, we identified its corresponding adversarial loss weighting that minimized structural disparity. Each optimally weighted adversarial loss yielded an average SSIM reduction of 1.5%. We further introduced a set of new metrics to assess other clinically relevant image features: Gradient Error (GE) to measure structural disparities; Sharpness to compute edge clarity; and Edge-Contrast Error (ECE) to quantify any distortion of the pixel distribution around edges. Including adversarial loss increased structural enhancement in visual inspection, which correlated with statistically consistent GE reductions (p-value << 0.05). This also resulted in increased Sharpness; however, the level of statistical significance was dependent on the perceptual loss weighting. Additionally, adversarial loss yielded ECE reductions for smaller perceptual loss weightings, while showing non-significant increases (p-value >> 0.05) when these weightings were higher, demonstrating that the increased Sharpness does not adversely distort the pixel distribution around the edges in the image. These studies clearly suggest that adversarial training significantly improves the performance of an MRI enhancement pipeline, and highlights the need for systematic studies of hyperparameter optimization and investigation of alternative image quality metrics.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    肌炎是肌肉的炎症,可能来自各种来源,具有不同的症状,需要不同的治疗方法。为了达到最佳治疗效果,及时获得准确的诊断至关重要。本文提出了一种新的监督分割架构,可以有效地执行精确的分割和分类从超声图像,很少的计算资源。我们模型的架构包括一个独特的编码器-解码器结构,该结构将瓶颈转换器(BOT)与新开发的名为Multi-ConvGhost可切换瓶颈残差块(MCG_RB)的残差块集成在一起。这个块以若干分辨率有效地捕获和分析编码器段内部的超声图像输入。此外,BOT模块是一个变压器式的注意模块,旨在弥合编码和解码阶段之间的功能差距。此外,使用MCG-RB模块检索多级特征,它将多卷积与卷积的鬼可切换残差连接相结合,用于编码和解码阶段。建议的方法在所有参数的一组基准肌炎超声图像上都达到了最先进的性能,包括准确性,精度,召回,骰子系数,和Jaccard指数。尽管训练数据有限,建议的方法通过产生出色的结果证明了显着的普适性。与Unet++等最先进的分割方法相比,所提出的模型在准确性上有了显著提高,DeepLabV3和Duck-Net。骰子系数和Jaccard指数获得了高达3%的改善,6%,7%,分别,超越其他方法。
    Myositis is the inflammation of the muscles that can arise from various sources with diverse symptoms and require different treatments. For treatment to achieve optimal results, it is essential to obtain an accurate diagnosis promptly. This paper presents a new supervised segmentation architecture that can efficiently perform precise segmentation and classification of myositis from ultrasound images with few computational resources. The architecture of our model includes a unique encoder-decoder structure that integrates the Bottleneck Transformer (BOT) with a newly developed Residual block named Multi-Conv Ghost switchable bottleneck Residual Block (MCG_RB). This block effectively captures and analyzes ultrasound image input inside the encoder segment at several resolutions. Moreover, the BOT module is a transformer-style attention module designed to bridge the feature gap between the encoding and decoding stages. Furthermore, multi-level features are retrieved using the MCG-RB module, which combines multi-convolution with ghost switchable residual connections of convolutions for both the encoding and decoding stages. The suggested method attains state-of-the-art performance on a benchmark set of myositis ultrasound images across all parameters, including accuracy, precision, recall, dice coefficient, and Jaccard index. Despite its limited training data, the suggested approach demonstrates remarkable generalizability by yielding exceptional results. The proposed model showed a substantial enhancement in accuracy when compared to segmentation state-of-the-art methods such as Unet++, DeepLabV3, and the Duck-Net. The dice coefficient and Jaccard index obtained improvements of up to 3%, 6%, and 7%, respectively, surpassing the other methods.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    背景:计算机断层扫描(CT)依赖于X射线的衰减,而且是,因此,对身体弱衰减器官的有限使用,比如肺。X射线暗场(DF)成像是最近开发的技术,其利用X射线光栅来实现小角度散射作为替代的对比度机制。DF信号提供有关物体微观形态的结构信息,补充传统的衰减信号。我们小组开发了第一个人体规模的X射线DFCT。尽管有专门的处理算法,重建的图像仍然受到条纹伪影的影响,这往往会阻碍图像的解释。近年来,卷积神经网络在CT重建领域得到了广泛的应用,除其他外,用于清除文物。
    目的:减少条纹伪影对于优化DFCT的图像质量至关重要,和无伪影图像是潜在的未来临床应用的先决条件。本文的目的是证明CNN后处理在X射线DFCT中减少伪影的可行性,以及多旋转扫描如何作为训练数据的途径。
    方法:我们采用了有监督的深度学习方法,使用三维双框架UNet来去除条纹伪影。所需的训练数据是从我们研究所的实验X射线DFCT原型获得的。使用两种不同的操作模式来生成输入和相应的地面实况数据集。剂量相容辐射水平的临床相关扫描被用作输入数据,和具有更少伪影的扩展扫描被用作地面实况数据。后者既不是剂量-,也不兼容时间,因此,对患者的临床成像不可行。
    结果:经过训练的CNN能够大大减少DFCT图像中的条纹伪影。对网络进行了完全不同的图像测试,以前看不见的图像特征。在所有情况下,CNN处理大大提高了图像质量,这通过增加的图像质量指标定量证实。在加工过程中保留了精细的细节,尽管输出图像看起来比地面实况图像更平滑。
    结论:我们的结果展示了神经网络减少X射线DFCT中条纹伪影的潜力。在剂量兼容的X射线DFCT中成功增强了图像质量,这对于X射线DFCT在现代临床放射学中的应用起着至关重要的作用。
    BACKGROUND: Computed tomography (CT) relies on the attenuation of x-rays, and is, hence, of limited use for weakly attenuating organs of the body, such as the lung. X-ray dark-field (DF) imaging is a recently developed technology that utilizes x-ray optical gratings to enable small-angle scattering as an alternative contrast mechanism. The DF signal provides structural information about the micromorphology of an object, complementary to the conventional attenuation signal. A first human-scale x-ray DF CT has been developed by our group. Despite specialized processing algorithms, reconstructed images remain affected by streaking artifacts, which often hinder image interpretation. In recent years, convolutional neural networks have gained popularity in the field of CT reconstruction, amongst others for streak artefact removal.
    OBJECTIVE: Reducing streak artifacts is essential for the optimization of image quality in DF CT, and artefact free images are a prerequisite for potential future clinical application. The purpose of this paper is to demonstrate the feasibility of CNN post-processing for artefact reduction in x-ray DF CT and how multi-rotation scans can serve as a pathway for training data.
    METHODS: We employed a supervised deep-learning approach using a three-dimensional dual-frame UNet in order to remove streak artifacts. Required training data were obtained from the experimental x-ray DF CT prototype at our institute. Two different operating modes were used to generate input and corresponding ground truth data sets. Clinically relevant scans at dose-compatible radiation levels were used as input data, and extended scans with substantially fewer artifacts were used as ground truth data. The latter is neither dose-, nor time-compatible and, therefore, unfeasible for clinical imaging of patients.
    RESULTS: The trained CNN was able to greatly reduce streak artifacts in DF CT images. The network was tested against images with entirely different, previously unseen image characteristics. In all cases, CNN processing substantially increased the image quality, which was quantitatively confirmed by increased image quality metrics. Fine details are preserved during processing, despite the output images appearing smoother than the ground truth images.
    CONCLUSIONS: Our results showcase the potential of a neural network to reduce streak artifacts in x-ray DF CT. The image quality is successfully enhanced in dose-compatible x-ray DF CT, which plays an essential role for the adoption of x-ray DF CT into modern clinical radiology.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    背景:MRI图像上的膀胱癌(BC)分割是确定是否存在肌肉浸润的第一步。本研究旨在评估三种深度学习(DL)模型在多参数MRI(mp-MRI)图像上的肿瘤分割性能。
    方法:我们研究了53例膀胱癌患者。膀胱肿瘤在T2加权(T2WI)的每个切片上进行分割,扩散加权成像/表观扩散系数(DWI/ADC),和在3TeslaMRI扫描仪上采集的T1加权对比增强(T1WI)图像。我们训练了Unet,MAnet,和PSPnet使用三个损失函数:交叉熵(CE),骰子相似系数损失(DSC),和病灶丢失(FL)。我们使用DSC评估了模型性能,Hausdorff距离(HD),和预期校准误差(ECE)。
    结果:具有CE+DSC损失函数的MAnet算法在ADC上给出了最高的DSC值,T2WI,和T1WI图像。PSPnet与CE+DSC在ADC上获得了最小的HDs,T2WI,和T1WI图像。总体上,ADC和T1WI的分割精度优于T2WI。在ADC图像上,带FL的PSPnet的ECE最小,而在T2WI和T1WI上使用CE+DSC的MAnet是最小的。
    结论:与Unet相比,根据评估指标的选择,具有混合CEDSC损失函数的MAnet和PSPnet在BC分割中显示出更好的性能。
    BACKGROUND: Bladder cancer (BC) segmentation on MRI images is the first step to determining the presence of muscular invasion. This study aimed to assess the tumor segmentation performance of three deep learning (DL) models on multi-parametric MRI (mp-MRI) images.
    METHODS: We studied 53 patients with bladder cancer. Bladder tumors were segmented on each slice of T2-weighted (T2WI), diffusion-weighted imaging/apparent diffusion coefficient (DWI/ADC), and T1-weighted contrast-enhanced (T1WI) images acquired at a 3Tesla MRI scanner. We trained Unet, MAnet, and PSPnet using three loss functions: cross-entropy (CE), dice similarity coefficient loss (DSC), and focal loss (FL). We evaluated the model performances using DSC, Hausdorff distance (HD), and expected calibration error (ECE).
    RESULTS: The MAnet algorithm with the CE+DSC loss function gave the highest DSC values on the ADC, T2WI, and T1WI images. PSPnet with CE+DSC obtained the smallest HDs on the ADC, T2WI, and T1WI images. The segmentation accuracy overall was better on the ADC and T1WI than on the T2WI. The ECEs were the smallest for PSPnet with FL on the ADC images, while they were the smallest for MAnet with CE+DSC on the T2WI and T1WI.
    CONCLUSIONS: Compared to Unet, MAnet and PSPnet with a hybrid CE+DSC loss function displayed better performances in BC segmentation depending on the choice of the evaluation metric.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    背景:本研究旨在开发一种称为双路径双注意转换器(DDA-Transformer)的新型深度卷积神经网络,旨在实现精确,快速的膝关节CT图像分割,并在机器人辅助的全膝关节置换术(TKA)中进行验证。
    方法:股骨,胫骨,髌骨,和腓骨分割的性能和速度进行了评估,组件尺寸的准确性,在临床上验证了使用该深度学习网络构建的机器人辅助TKA系统的骨切除和对齐.
    结果:总体而言,DDA-Transformer在Dice系数方面优于其他六个网络,在联合上相交,平均表面距离,和Hausdorff距离.DDA-Transformer的分割速度明显快于nnUnet,TransUnet和3D-Unet(p<0.01)。此外,机器人辅助TKA系统在手术准确性方面优于手动组.
    结论:DDA-Transformer在膝关节分割中显示出显着提高的准确性和鲁棒性,这种方便稳定的膝关节CT图像分割网络显著提高了TKA程序的准确性。
    BACKGROUND: This study aimed to develop a novel deep convolutional neural network called Dual-path Double Attention Transformer (DDA-Transformer) designed to achieve precise and fast knee joint CT image segmentation and to validate it in robotic-assisted total knee arthroplasty (TKA).
    METHODS: The femoral, tibial, patellar, and fibular segmentation performance and speed were evaluated and the accuracy of component sizing, bone resection and alignment of the robotic-assisted TKA system constructed using this deep learning network was clinically validated.
    RESULTS: Overall, DDA-Transformer outperformed six other networks in terms of the Dice coefficient, intersection over union, average surface distance, and Hausdorff distance. DDA-Transformer exhibited significantly faster segmentation speeds than nnUnet, TransUnet and 3D-Unet (p < 0.01). Furthermore, the robotic-assisted TKA system outperforms the manual group in surgical accuracy.
    CONCLUSIONS: DDA-Transformer exhibited significantly improved accuracy and robustness in knee joint segmentation, and this convenient and stable knee joint CT image segmentation network significantly improved the accuracy of the TKA procedure.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    蛋白质对生命至关重要,理解它们的内在角色需要确定它们的结构。通过将深度学习算法应用于已解决的蛋白质结构的大型数据库,蛋白质组学领域开辟了新的机遇。随着大型数据集和先进的机器学习方法的可用性,蛋白质残基相互作用的预测有了很大的提高。蛋白质接触图提供了蛋白质序列内相互作用的残基对的经验证据。无模板的蛋白质结构预测系统严重依赖于这些信息。本文提出了UNet-CON,注意综合的UNet架构,训练预测蛋白质序列中的残基-残基接触。由于预测的接触比PDB25测试装置上的最新方法更准确,该模型为预测蛋白质残基相互作用的更强大的深度学习算法的发展铺平了道路。源代码可在GitHub链接中找到:(https://github.com/jisnava/UNetCON)。
    Proteins are essential to life, and understanding their intrinsic roles requires determining their structure. The field of proteomics has opened up new opportunities by applying deep learning algorithms to large databases of solved protein structures. With the availability of large data sets and advanced machine learning methods, the prediction of protein residue interactions has greatly improved. Protein contact maps provide empirical evidence of the interacting residue pairs within a protein sequence. Template-free protein structure prediction systems rely heavily on this information. This article proposes UNet-CON, an attention-integrated UNet architecture, trained to predict residue-residue contacts in protein sequences. With the predicted contacts being more accurate than state-of-the-art methods on the PDB25 test set, the model paves the way for the development of more powerful deep learning algorithms for predicting protein residue interactions.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    神经胶质瘤脑图像中肿瘤区域的迅速和精确的识别和描绘对于减轻与这种危及生命的疾病相关的风险至关重要。在这项研究中,我们采用UNet卷积神经网络(CNN)架构进行胶质瘤肿瘤检测.我们提出的方法包括一个转换模块,特征提取模块,和肿瘤分割模块。脑磁共振成像图像的空间域表示通过非子采样Shearlet变换分解为低频和高频子带。利用此转换的选择性和指导性特征增强了我们提出的系统的分类功效。从低频和高频子带提取剪切特征,随后使用UNet-CNN架构进行分类,以识别神经胶质瘤脑图像中的肿瘤区域。我们使用公开可用的数据集验证我们提出的神经胶质瘤肿瘤检测方法,即脑肿瘤分割(BRATS)2019和癌症基因组图谱(TCGA)。我们的系统实现的平均分类率在BRATS2019数据集为99.1%,在TCGA数据集为97.8%。此外,我们的系统在BRATS2019数据集上展示了显著的性能指标,包括98.2%的灵敏度,98.7%的特异性,98.9%的准确度,98.7%的交叉超过并网,和98.5%的圆盘相似系数。同样,在TCGA数据集上,我们的系统达到97.7%的灵敏度,98.2%的特异性,98.7%的准确度,98.6%的交叉超过工会,椎间盘相似系数为98.4%。与最新方法的比较分析强调了我们提出的神经胶质瘤脑肿瘤检测方法的有效性。
    The prompt and precise identification and delineation of tumor regions within glioma brain images are critical for mitigating the risks associated with this life-threatening ailment. In this study, we employ the UNet convolutional neural network (CNN) architecture for glioma tumor detection. Our proposed methodology comprises a transformation module, a feature extraction module, and a tumor segmentation module. The spatial domain representation of brain magnetic resonance imaging images undergoes decomposition into low- and high-frequency subbands via a non-subsampled shearlet transform. Leveraging the selective and directive characteristics of this transform enhances the classification efficacy of our proposed system. Shearlet features are extracted from both low- and high-frequency subbands and subsequently classified using the UNet-CNN architecture to identify tumor regions within glioma brain images. We validate our proposed glioma tumor detection methodology using publicly available datasets, namely Brain Tumor Segmentation (BRATS) 2019 and The Cancer Genome Atlas (TCGA). The mean classification rates achieved by our system are 99.1% for the BRATS 2019 dataset and 97.8% for the TCGA dataset. Furthermore, our system demonstrates notable performance metrics on the BRATS 2019 dataset, including 98.2% sensitivity, 98.7% specificity, 98.9% accuracy, 98.7% intersection over union, and 98.5% disc similarity coefficient. Similarly, on the TCGA dataset, our system achieves 97.7% sensitivity, 98.2% specificity, 98.7% accuracy, 98.6% intersection over union, and 98.4% disc similarity coefficient. Comparative analysis against state-of-the-art methods underscores the efficacy of our proposed glioma brain tumor detection approach.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    在过去的二十年里,医学影像的机器分析发展迅速,为几个重要的医疗应用开辟了巨大的潜力。随着复杂疾病的增加和病例数的增加,基于机器的成像分析的作用已经变得不可或缺。它既是医学专家的工具,也是医学专家的助手,提供有价值的见解和指导。在这个领域一个特别具有挑战性的任务是病变分割,即使对于经验丰富的放射科医生来说,这项任务也具有挑战性。这项任务的复杂性凸显了迫切需要强大的机器学习方法来支持医务人员。作为回应,我们提出了我们的新解决方案:D-TrAttUnet体系结构。该框架基于不同疾病通常靶向特定器官的观察。我们的架构包括具有复合Transformer-CNN编码器和双解码器的编码器-解码器结构。编码器包括两个路径:变换器路径和编码器融合模块路径。双解码器配置使用两个相同的解码器,每个人都有注意门。这允许模型同时分割病变和器官并整合它们的分割损失。为了验证我们的方法,我们对Covid-19和骨转移分割任务进行了评估。我们还通过在没有第二个解码器的情况下对腺体和细胞核的分割进行测试来研究模型的适应性。结果证实了我们方法的优越性,特别是在新冠肺炎感染和骨转移的分割中。此外,混合编码器在腺体和细胞核的分割中表现出卓越的性能,巩固其在现代医学图像分析中的作用。
    Over the past two decades, machine analysis of medical imaging has advanced rapidly, opening up significant potential for several important medical applications. As complicated diseases increase and the number of cases rises, the role of machine-based imaging analysis has become indispensable. It serves as both a tool and an assistant to medical experts, providing valuable insights and guidance. A particularly challenging task in this area is lesion segmentation, a task that is challenging even for experienced radiologists. The complexity of this task highlights the urgent need for robust machine learning approaches to support medical staff. In response, we present our novel solution: the D-TrAttUnet architecture. This framework is based on the observation that different diseases often target specific organs. Our architecture includes an encoder-decoder structure with a composite Transformer-CNN encoder and dual decoders. The encoder includes two paths: the Transformer path and the Encoders Fusion Module path. The Dual-Decoder configuration uses two identical decoders, each with attention gates. This allows the model to simultaneously segment lesions and organs and integrate their segmentation losses. To validate our approach, we performed evaluations on the Covid-19 and Bone Metastasis segmentation tasks. We also investigated the adaptability of the model by testing it without the second decoder in the segmentation of glands and nuclei. The results confirmed the superiority of our approach, especially in Covid-19 infections and the segmentation of bone metastases. In addition, the hybrid encoder showed exceptional performance in the segmentation of glands and nuclei, solidifying its role in modern medical image analysis.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

公众号