Medical image processing

医学图像处理
  • 文章类型: Journal Article
    冠状动脉疾病仍然是心血管疾病患者死亡的主要原因。通过支架植入的生物可吸收血管支架(BVS)的治疗用途很常见,然而,目前来自血管内光学相干断层扫描(IVOCT)图像的BVS分割技术的有效性不足.
    本文介绍了一种增强的分割方法,该方法使用一种新颖的基于小波的U形网络来解决这些挑战。我们开发了一种基于小波的U形网络,该网络结合了注意力门(AG)和Atrous多尺度场模块(AMFM),旨在通过改善支架支柱与周围组织之间的差异来提高分割精度。独特的小波融合模块缓解了不同特征图分支之间的语义差距,促进更有效的功能集成。
    大量实验表明,我们的模型在关键指标如Dice系数方面超越了现有技术,准确度,灵敏度,和联合交汇处(IoU),得分达到85.10%,99.77%,86.93%,73.81%,分别。AG的整合,AMFM,融合模块在实现这些结果方面发挥了关键作用,指示在捕获详细的上下文信息方面的显著增强。
    基于小波的U形网络的引入标志着IVOCT图像中BVSs分割的实质性改进,提示冠心病治疗临床实践的潜在益处。这种方法也可能适用于其他复杂的医学成像分割任务,表明未来研究的广阔范围。
    UNASSIGNED: Coronary artery disease remains a leading cause of mortality among individuals with cardiovascular conditions. The therapeutic use of bioresorbable vascular scaffolds (BVSs) through stent implantation is common, yet the effectiveness of current BVS segmentation techniques from Intravascular Optical Coherence Tomography (IVOCT) images is inadequate.
    UNASSIGNED: This paper introduces an enhanced segmentation approach using a novel Wavelet-based U-shape network to address these challenges. We developed a Wavelet-based U-shape network that incorporates an Attention Gate (AG) and an Atrous Multi-scale Field Module (AMFM), designed to enhance the segmentation accuracy by improving the differentiation between the stent struts and the surrounding tissue. A unique wavelet fusion module mitigates the semantic gaps between different feature map branches, facilitating more effective feature integration.
    UNASSIGNED: Extensive experiments demonstrate that our model surpasses existing techniques in key metrics such as Dice coefficient, accuracy, sensitivity, and Intersection over Union (IoU), achieving scores of 85.10%, 99.77%, 86.93%, and 73.81%, respectively. The integration of AG, AMFM, and the fusion module played a crucial role in achieving these outcomes, indicating a significant enhancement in capturing detailed contextual information.
    UNASSIGNED: The introduction of the Wavelet-based U-shape network marks a substantial improvement in the segmentation of BVSs in IVOCT images, suggesting potential benefits for clinical practices in coronary artery disease treatment. This approach may also be applicable to other intricate medical imaging segmentation tasks, indicating a broad scope for future research.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    随着注意力等机制的引入,现有病变语义分割模型的性能已显示出稳步改善,跳过连接,深度监督。然而,这些进步通常以计算需求为代价,需要具有大量视频内存的强大图形处理单元。因此,某些型号在更实惠的边缘设备上可能表现出较差或不存在的性能,例如智能手机和其他即时设备。为了应对这一挑战,本文介绍了一种低参数计数和最小操作的病变分割模型。该模型结合了极坐标变换来简化图像,促进更快的培训和提高性能。我们通过将模型的焦点引导到最可能包含分割信息的区域来利用极点图像的特征,通过引入一种学习高效的基于极坐标的对比注意力(PCA)来实现。此设计利用Hadamard产品来实现轻量级注意力机制,而不会显着增加模型参数和复杂性。此外,我们提出了一种新的跳过跨通道聚合(SC2A)方法,用于共享跨通道校正,引入高斯深度卷积来增强非线性。对ISIC2018和Kvasir数据集的大量实验表明,我们的模型超越了最先进的模型,同时只保留了大约25K个参数。此外,我们提出的模型对跨域数据表现出很强的泛化能力,通过PH2数据集和CVC-Polyp数据集的实验证实。此外,我们评估模型的性能在移动设置与其他轻量级模型。值得注意的是,我们提出的模型在IoU和Dice评分方面优于其他高级模型,运行时间。
    The performance of existing lesion semantic segmentation models has shown a steady improvement with the introduction of mechanisms like attention, skip connections, and deep supervision. However, these advancements often come at the expense of computational requirements, necessitating powerful graphics processing units with substantial video memory. Consequently, certain models may exhibit poor or non-existent performance on more affordable edge devices, such as smartphones and other point-of-care devices. To tackle this challenge, our paper introduces a lesion segmentation model with a low parameter count and minimal operations. This model incorporates polar transformations to simplify images, facilitating faster training and improved performance. We leverage the characteristics of polar images by directing the model\'s focus to areas most likely to contain segmentation information, achieved through the introduction of a learning-efficient polar-based contrast attention (PCA). This design utilizes Hadamard products to implement a lightweight attention mechanism without significantly increasing model parameters and complexities. Furthermore, we present a novel skip cross-channel aggregation (SC2A) approach for sharing cross-channel corrections, introducing Gaussian depthwise convolution to enhance nonlinearity. Extensive experiments on the ISIC 2018 and Kvasir datasets demonstrate that our model surpasses state-of-the-art models while maintaining only about 25K parameters. Additionally, our proposed model exhibits strong generalization to cross-domain data, as confirmed through experiments on the PH2 dataset and CVC-Polyp dataset. In addition, we evaluate the model\'s performance in a mobile setting against other lightweight models. Notably, our proposed model outperforms other advanced models in terms of IoU and Dice score, and running time.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    背景:多种疾病的胸部X线图像分类是计算机视觉和医学图像处理领域的重要研究方向。它旨在利用先进的图像处理技术和深度学习算法来自动分析和识别X射线图像,确定图像中是否存在特定的病理或结构异常。
    目的:我们提出了专为胸部多标签疾病分类而设计的MMPDenseNet网络。
    方法:最初,网络采用自适应激活函数Meta-ACON来增强特征表示。随后,该网络包含多头自我注意机制,将传统的卷积神经网络与Transformer合并,从而增强提取局部和全局特征的能力。最终,该网络集成了金字塔挤压注意力模块,以捕获空间信息并丰富特征空间。
    结果:结论实验产生的平均AUC为0.898,与基线模型相比,平均准确度提高了0.6%。与原始网络相比,实验结果表明,MMPDenseNet大大提高了各种胸部疾病的分类精度。
    结论:可以得出结论,因此,具有重要的临床应用价值。
    BACKGROUND: Chest X-ray image classification for multiple diseases is an important research direction in the field of computer vision and medical image processing. It aims to utilize advanced image processing techniques and deep learning algorithms to automatically analyze and identify X-ray images, determining whether specific pathologies or structural abnormalities exist in the images.
    OBJECTIVE: We present the MMPDenseNet network designed specifically for chest multi-label disease classification.
    METHODS: Initially, the network employs the adaptive activation function Meta-ACON to enhance feature representation. Subsequently, the network incorporates a multi-head self-attention mechanism, merging the conventional convolutional neural network with the Transformer, thereby bolstering the ability to extract both local and global features. Ultimately, the network integrates a pyramid squeeze attention module to capture spatial information and enrich the feature space.
    RESULTS: The concluding experiment yielded an average AUC of 0.898, marking an average accuracy improvement of 0.6% over the baseline model. When compared with the original network, the experimental results highlight that MMPDenseNet considerably elevates the classification accuracy of various chest diseases.
    CONCLUSIONS: It can be concluded that the network, thus, holds substantial value for clinical applications.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    近年来,手术机器人在微创外科领域的应用发展迅速,受到越来越多的研究关注。人们已经达成共识,即外科手术将减少创伤,并实施更多的智慧和更高的自主性,这是机器人系统环境感知能力面临的严峻挑战。机器人环境信息的主要来源之一是图像,这是机器人视觉的基础。在这篇评论文章中,根据信息获取的对象将临床图像分为直接图像和间接图像,并成为连续的,间歇连续,并且根据目标跟踪频率不连续。基于这两个维度介绍了现有手术机器人在各个范畴的特点和应用。我们进行这次审查的目的是分析,总结,并讨论当前关于医学应用图像技术的一般规则的证据。我们的分析提供了见解,并为将来开发更先进的手术机器人系统提供了指导。
    Surgical robotics application in the field of minimally invasive surgery has developed rapidly and has been attracting increasingly more research attention in recent years. A common consensus has been reached that surgical procedures are to become less traumatic and with the implementation of more intelligence and higher autonomy, which is a serious challenge faced by the environmental sensing capabilities of robotic systems. One of the main sources of environmental information for robots are images, which are the basis of robot vision. In this review article, we divide clinical image into direct and indirect based on the object of information acquisition, and into continuous, intermittent continuous, and discontinuous according to the target-tracking frequency. The characteristics and applications of the existing surgical robots in each category are introduced based on these two dimensions. Our purpose in conducting this review was to analyze, summarize, and discuss the current evidence on the general rules on the application of image technologies for medical purposes. Our analysis gives insight and provides guidance conducive to the development of more advanced surgical robotics systems in the future.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    高光谱成像已经证明了其通过非接触和非侵入性技术提供样本的相关空间和光谱信息的潜力。在医学领域,尤其是在组织病理学方面,HSI已用于病变组织的分类和鉴定以及其形态特性的表征。在这项工作中,我们提出了一种混合方案,通过高光谱成像对非肿瘤和肿瘤组织学脑样本进行分类。所提出的方法基于通过线性解混识别高光谱图像中的特征成分,作为一个特征工程步骤,并通过深度学习方法进行后续分类。这最后一步,通过增强数据集上的交叉验证方案和迁移学习方案来评估深度神经网络的集合。所提出的方法可以对组织学脑样本进行分类,平均准确率为88%,减少可变性,计算成本,和推理时间,这与最先进的方法相比具有优势。因此,这项工作证明了混合分类方法通过结合用于特征提取的线性分解和用于分类的深度学习来实现稳健和可靠的结果的潜力。
    Hyperspectral imaging has demonstrated its potential to provide correlated spatial and spectral information of a sample by a non-contact and non-invasive technology. In the medical field, especially in histopathology, HSI has been applied for the classification and identification of diseased tissue and for the characterization of its morphological properties. In this work, we propose a hybrid scheme to classify non-tumor and tumor histological brain samples by hyperspectral imaging. The proposed approach is based on the identification of characteristic components in a hyperspectral image by linear unmixing, as a features engineering step, and the subsequent classification by a deep learning approach. For this last step, an ensemble of deep neural networks is evaluated by a cross-validation scheme on an augmented dataset and a transfer learning scheme. The proposed method can classify histological brain samples with an average accuracy of 88%, and reduced variability, computational cost, and inference times, which presents an advantage over methods in the state-of-the-art. Hence, the work demonstrates the potential of hybrid classification methodologies to achieve robust and reliable results by combining linear unmixing for features extraction and deep learning for classification.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    脑肿瘤是由于异常细胞组织的扩张而发生的,可以是恶性的(癌性的)或良性的(非癌性的)。位置等众多因素,尺寸,在检测和诊断脑肿瘤时考虑进展率。在初始阶段检测脑肿瘤对于MRI(磁共振成像)扫描起着重要作用的诊断至关重要。多年来,深度学习模型已被广泛用于医学图像处理。目前的研究主要调查了新颖的微调视觉变换器模型(FTVT)-FTVT-b16,FTVT-b32,FTVT-l16,FTVT-l32-用于脑肿瘤分类,同时还将它们与其他已建立的深度学习模型进行比较,例如ResNet50、MobileNet-V2和EfficientNet-B0。包含7,023张图像(MRI扫描)的数据集分为四个不同的类别,即,神经胶质瘤,脑膜瘤,垂体,并且没有肿瘤用于分类。Further,该研究对这些模型进行了比较分析,包括它们的准确性和其他评估指标,包括召回,精度,每个班级的F1得分。深度学习模型ResNet-50、EfficientNet-B0和MobileNet-V2的准确率为96.5%,95.1%,94.9%,分别。在所有的FTVT模型中,FTVT-l16模型取得了98.70%的显著精度,而其他FTVT-b16、FTVT-b32和FTVT-132模型取得了98.09%的精度,96.87%,98.62%,分别,从而证明了FTVT在医学图像处理中的有效性和鲁棒性。
    Brain tumors occur due to the expansion of abnormal cell tissues and can be malignant (cancerous) or benign (not cancerous). Numerous factors such as the position, size, and progression rate are considered while detecting and diagnosing brain tumors. Detecting brain tumors in their initial phases is vital for diagnosis where MRI (magnetic resonance imaging) scans play an important role. Over the years, deep learning models have been extensively used for medical image processing. The current study primarily investigates the novel Fine-Tuned Vision Transformer models (FTVTs)-FTVT-b16, FTVT-b32, FTVT-l16, FTVT-l32-for brain tumor classification, while also comparing them with other established deep learning models such as ResNet50, MobileNet-V2, and EfficientNet - B0. A dataset with 7,023 images (MRI scans) categorized into four different classes, namely, glioma, meningioma, pituitary, and no tumor are used for classification. Further, the study presents a comparative analysis of these models including their accuracies and other evaluation metrics including recall, precision, and F1-score across each class. The deep learning models ResNet-50, EfficientNet-B0, and MobileNet-V2 obtained an accuracy of 96.5%, 95.1%, and 94.9%, respectively. Among all the FTVT models, FTVT-l16 model achieved a remarkable accuracy of 98.70% whereas other FTVT models FTVT-b16, FTVT-b32, and FTVT-132 achieved an accuracy of 98.09%, 96.87%, 98.62%, respectively, hence proving the efficacy and robustness of FTVT\'s in medical image processing.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    将自动分割方法结合到牙科X射线图像中,通过促进细致,完善了临床诊断和治疗计划的范例,牙齿结构和邻近组织的像素级关节。这是早期病理检测和细致的疾病进展监测的支柱。尽管如此,由于X射线成像的内在局限性,传统的分割框架经常会遇到重大挫折,包括受损的图像保真度,结构边界的模糊划定,以及牙髓等牙齿成分的复杂解剖结构,搪瓷,还有牙本质.为了克服这些障碍,我们提出了可变形卷积和Mamba集成网络,创新的2D牙科X射线图像分割架构,合并了一个合并结构可变形编码器,认知优化的语义增强模块,和分层收敛解码器。总的来说,这些组件支持多尺度全球功能的管理,加强特征表示的稳定性,并完善特征向量的合并。对14个基线的比较评估强调了其有效性,记录骰子系数增加了0.95%,第95个百分位数Hausdorff距离减少到7.494。
    The incorporation of automatic segmentation methodologies into dental X-ray images refined the paradigms of clinical diagnostics and therapeutic planning by facilitating meticulous, pixel-level articulation of both dental structures and proximate tissues. This underpins the pillars of early pathological detection and meticulous disease progression monitoring. Nonetheless, conventional segmentation frameworks often encounter significant setbacks attributable to the intrinsic limitations of X-ray imaging, including compromised image fidelity, obscured delineation of structural boundaries, and the intricate anatomical structures of dental constituents such as pulp, enamel, and dentin. To surmount these impediments, we propose the Deformable Convolution and Mamba Integration Network, an innovative 2D dental X-ray image segmentation architecture, which amalgamates a Coalescent Structural Deformable Encoder, a Cognitively-Optimized Semantic Enhance Module, and a Hierarchical Convergence Decoder. Collectively, these components bolster the management of multi-scale global features, fortify the stability of feature representation, and refine the amalgamation of feature vectors. A comparative assessment against 14 baselines underscores its efficacy, registering a 0.95% enhancement in the Dice Coefficient and a diminution of the 95th percentile Hausdorff Distance to 7.494.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    在受设备限制的临床条件下,实现轻量级的皮肤病变分割至关重要,因为它有助于将模型集成到各种医疗设备中,从而提高运营效率。然而,模型的轻量化设计可能面临精度下降,特别是当处理复杂的图像,如皮肤病变图像与不规则区域,模糊的边界,和超大的边界。为了应对这些挑战,我们提出了一个有效的轻量级注意网络(ELANet)用于皮肤病变分割任务。在ELANet,两种不同的注意机制的双边残差模块(BRM)可以实现信息互补,这增强了对空间和通道维度特征的敏感性,分别,然后将多个BRM堆叠起来,对输入信息进行有效的特征提取。此外,该网络通过多尺度注意力融合(MAF)操作放置不同尺度的特征图来获取全局信息并提高分割精度。最后,我们评估了ELANet在三个公开可用数据集上的性能,ISIC2016、ISIC2017和ISIC2018,实验结果表明,我们的算法可以达到89.87%,81.85%,三个参数为0.459M的数据集上的mIoU的82.87%,这是一个很好的平衡之间的准确性和亮度,是优于许多现有的分割方法。
    In clinical conditions limited by equipment, attaining lightweight skin lesion segmentation is pivotal as it facilitates the integration of the model into diverse medical devices, thereby enhancing operational efficiency. However, the lightweight design of the model may face accuracy degradation, especially when dealing with complex images such as skin lesion images with irregular regions, blurred boundaries, and oversized boundaries. To address these challenges, we propose an efficient lightweight attention network (ELANet) for the skin lesion segmentation task. In ELANet, two different attention mechanisms of the bilateral residual module (BRM) can achieve complementary information, which enhances the sensitivity to features in spatial and channel dimensions, respectively, and then multiple BRMs are stacked for efficient feature extraction of the input information. In addition, the network acquires global information and improves segmentation accuracy by putting feature maps of different scales through multi-scale attention fusion (MAF) operations. Finally, we evaluate the performance of ELANet on three publicly available datasets, ISIC2016, ISIC2017, and ISIC2018, and the experimental results show that our algorithm can achieve 89.87%, 81.85%, and 82.87% of the mIoU on the three datasets with a parametric of 0.459 M, which is an excellent balance between accuracy and lightness and is superior to many existing segmentation methods.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    骨小梁分析在了解骨骼健康和疾病中起着至关重要的作用。应用像骨质疏松症诊断。本文对三维骨小梁CT图像复原进行了全面的研究,解决这一领域的重大挑战。这项研究引入了一个骨干模型,级联-SwinUNETR,单视图三维CT图像复原。该模型利用具有监督和Swin-Transformer功能的深层聚合,在特征提取方面表现出色。此外,这项研究还带来了DVSR3D,双视图恢复模型,通过深度特征融合与注意力机制和自动编码器实现良好的性能。此外,介绍了一种无监督域自适应(UDA)方法,允许模型在没有额外标签的情况下适应输入数据分布,在现实世界的医疗应用中拥有巨大的潜力,并消除了对侵入性数据收集程序的需求。该研究还包括用于CT图像复原的新的双视图数据集的策展,解决Micro-CT中真实人体骨骼数据的稀缺性。最后,通过下游医学骨微结构测量验证了双视图方法。我们的贡献为骨小梁分析开辟了几条途径,有望改善骨健康评估和诊断的临床结果。
    Trabecular bone analysis plays a crucial role in understanding bone health and disease, with applications like osteoporosis diagnosis. This paper presents a comprehensive study on 3D trabecular computed tomography (CT) image restoration, addressing significant challenges in this domain. The research introduces a backbone model, Cascade-SwinUNETR, for single-view 3D CT image restoration. This model leverages deep layer aggregation with supervision and capabilities of Swin-Transformer to excel in feature extraction. Additionally, this study also brings DVSR3D, a dual-view restoration model, achieving good performance through deep feature fusion with attention mechanisms and Autoencoders. Furthermore, an Unsupervised Domain Adaptation (UDA) method is introduced, allowing models to adapt to input data distributions without additional labels, holding significant potential for real-world medical applications, and eliminating the need for invasive data collection procedures. The study also includes the curation of a new dual-view dataset for CT image restoration, addressing the scarcity of real human bone data in Micro-CT. Finally, the dual-view approach is validated through downstream medical bone microstructure measurements. Our contributions open several paths for trabecular bone analysis, promising improved clinical outcomes in bone health assessment and diagnosis.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    医学图像分割通常涉及多种组织类型和结构,包括血管分割和神经纤维束分割等任务。增强分割结果的连续性是医学图像分割的关键挑战,在临床应用需求的推动下,专注于疾病的定位和量化。在这项研究中,一种新颖的分割模型是专门为视网膜血管分割设计的,利用船只方位信息,边界约束,和连续性约束,以提高分割精度。为了实现这一点,我们将U-Net与长短期记忆网络(LSTM)级联。U-Net的特点是参数数量少,分割效率高,而LSTM提供参数共享功能。此外,我们引入了一个方向信息增强模块插入到模型的底层,通过方向卷积算子获得包含方向信息的特征图。此外,我们设计了一个新的混合损失函数,它由连接损失组成,边界损失,和交叉熵损失。实验结果表明,该模型在三个广泛认可的视网膜血管分割数据集上实现了出色的分割结果,CHASE_DB1,DRIVE,还有ARIA.
    Medical image segmentation commonly involves diverse tissue types and structures, including tasks such as blood vessel segmentation and nerve fiber bundle segmentation. Enhancing the continuity of segmentation outcomes represents a pivotal challenge in medical image segmentation, driven by the demands of clinical applications, focusing on disease localization and quantification. In this study, a novel segmentation model is specifically designed for retinal vessel segmentation, leveraging vessel orientation information, boundary constraints, and continuity constraints to improve segmentation accuracy. To achieve this, we cascade U-Net with a long-short-term memory network (LSTM). U-Net is characterized by a small number of parameters and high segmentation efficiency, while LSTM offers a parameter-sharing capability. Additionally, we introduce an orientation information enhancement module inserted into the model\'s bottom layer to obtain feature maps containing orientation information through an orientation convolution operator. Furthermore, we design a new hybrid loss function that consists of connectivity loss, boundary loss, and cross-entropy loss. Experimental results demonstrate that the model achieves excellent segmentation outcomes across three widely recognized retinal vessel segmentation datasets, CHASE_DB1, DRIVE, and ARIA.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

公众号