Medical image processing

医学图像处理
  • 文章类型: Journal Article
    背景:多种疾病的胸部X线图像分类是计算机视觉和医学图像处理领域的重要研究方向。它旨在利用先进的图像处理技术和深度学习算法来自动分析和识别X射线图像,确定图像中是否存在特定的病理或结构异常。
    目的:我们提出了专为胸部多标签疾病分类而设计的MMPDenseNet网络。
    方法:最初,网络采用自适应激活函数Meta-ACON来增强特征表示。随后,该网络包含多头自我注意机制,将传统的卷积神经网络与Transformer合并,从而增强提取局部和全局特征的能力。最终,该网络集成了金字塔挤压注意力模块,以捕获空间信息并丰富特征空间。
    结果:结论实验产生的平均AUC为0.898,与基线模型相比,平均准确度提高了0.6%。与原始网络相比,实验结果表明,MMPDenseNet大大提高了各种胸部疾病的分类精度。
    结论:可以得出结论,因此,具有重要的临床应用价值。
    BACKGROUND: Chest X-ray image classification for multiple diseases is an important research direction in the field of computer vision and medical image processing. It aims to utilize advanced image processing techniques and deep learning algorithms to automatically analyze and identify X-ray images, determining whether specific pathologies or structural abnormalities exist in the images.
    OBJECTIVE: We present the MMPDenseNet network designed specifically for chest multi-label disease classification.
    METHODS: Initially, the network employs the adaptive activation function Meta-ACON to enhance feature representation. Subsequently, the network incorporates a multi-head self-attention mechanism, merging the conventional convolutional neural network with the Transformer, thereby bolstering the ability to extract both local and global features. Ultimately, the network integrates a pyramid squeeze attention module to capture spatial information and enrich the feature space.
    RESULTS: The concluding experiment yielded an average AUC of 0.898, marking an average accuracy improvement of 0.6% over the baseline model. When compared with the original network, the experimental results highlight that MMPDenseNet considerably elevates the classification accuracy of various chest diseases.
    CONCLUSIONS: It can be concluded that the network, thus, holds substantial value for clinical applications.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    近年来,手术机器人在微创外科领域的应用发展迅速,受到越来越多的研究关注。人们已经达成共识,即外科手术将减少创伤,并实施更多的智慧和更高的自主性,这是机器人系统环境感知能力面临的严峻挑战。机器人环境信息的主要来源之一是图像,这是机器人视觉的基础。在这篇评论文章中,根据信息获取的对象将临床图像分为直接图像和间接图像,并成为连续的,间歇连续,并且根据目标跟踪频率不连续。基于这两个维度介绍了现有手术机器人在各个范畴的特点和应用。我们进行这次审查的目的是分析,总结,并讨论当前关于医学应用图像技术的一般规则的证据。我们的分析提供了见解,并为将来开发更先进的手术机器人系统提供了指导。
    Surgical robotics application in the field of minimally invasive surgery has developed rapidly and has been attracting increasingly more research attention in recent years. A common consensus has been reached that surgical procedures are to become less traumatic and with the implementation of more intelligence and higher autonomy, which is a serious challenge faced by the environmental sensing capabilities of robotic systems. One of the main sources of environmental information for robots are images, which are the basis of robot vision. In this review article, we divide clinical image into direct and indirect based on the object of information acquisition, and into continuous, intermittent continuous, and discontinuous according to the target-tracking frequency. The characteristics and applications of the existing surgical robots in each category are introduced based on these two dimensions. Our purpose in conducting this review was to analyze, summarize, and discuss the current evidence on the general rules on the application of image technologies for medical purposes. Our analysis gives insight and provides guidance conducive to the development of more advanced surgical robotics systems in the future.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    高光谱成像已经证明了其通过非接触和非侵入性技术提供样本的相关空间和光谱信息的潜力。在医学领域,尤其是在组织病理学方面,HSI已用于病变组织的分类和鉴定以及其形态特性的表征。在这项工作中,我们提出了一种混合方案,通过高光谱成像对非肿瘤和肿瘤组织学脑样本进行分类。所提出的方法基于通过线性解混识别高光谱图像中的特征成分,作为一个特征工程步骤,并通过深度学习方法进行后续分类。这最后一步,通过增强数据集上的交叉验证方案和迁移学习方案来评估深度神经网络的集合。所提出的方法可以对组织学脑样本进行分类,平均准确率为88%,减少可变性,计算成本,和推理时间,这与最先进的方法相比具有优势。因此,这项工作证明了混合分类方法通过结合用于特征提取的线性分解和用于分类的深度学习来实现稳健和可靠的结果的潜力。
    Hyperspectral imaging has demonstrated its potential to provide correlated spatial and spectral information of a sample by a non-contact and non-invasive technology. In the medical field, especially in histopathology, HSI has been applied for the classification and identification of diseased tissue and for the characterization of its morphological properties. In this work, we propose a hybrid scheme to classify non-tumor and tumor histological brain samples by hyperspectral imaging. The proposed approach is based on the identification of characteristic components in a hyperspectral image by linear unmixing, as a features engineering step, and the subsequent classification by a deep learning approach. For this last step, an ensemble of deep neural networks is evaluated by a cross-validation scheme on an augmented dataset and a transfer learning scheme. The proposed method can classify histological brain samples with an average accuracy of 88%, and reduced variability, computational cost, and inference times, which presents an advantage over methods in the state-of-the-art. Hence, the work demonstrates the potential of hybrid classification methodologies to achieve robust and reliable results by combining linear unmixing for features extraction and deep learning for classification.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    脑肿瘤是由于异常细胞组织的扩张而发生的,可以是恶性的(癌性的)或良性的(非癌性的)。位置等众多因素,尺寸,在检测和诊断脑肿瘤时考虑进展率。在初始阶段检测脑肿瘤对于MRI(磁共振成像)扫描起着重要作用的诊断至关重要。多年来,深度学习模型已被广泛用于医学图像处理。目前的研究主要调查了新颖的微调视觉变换器模型(FTVT)-FTVT-b16,FTVT-b32,FTVT-l16,FTVT-l32-用于脑肿瘤分类,同时还将它们与其他已建立的深度学习模型进行比较,例如ResNet50、MobileNet-V2和EfficientNet-B0。包含7,023张图像(MRI扫描)的数据集分为四个不同的类别,即,神经胶质瘤,脑膜瘤,垂体,并且没有肿瘤用于分类。Further,该研究对这些模型进行了比较分析,包括它们的准确性和其他评估指标,包括召回,精度,每个班级的F1得分。深度学习模型ResNet-50、EfficientNet-B0和MobileNet-V2的准确率为96.5%,95.1%,94.9%,分别。在所有的FTVT模型中,FTVT-l16模型取得了98.70%的显著精度,而其他FTVT-b16、FTVT-b32和FTVT-132模型取得了98.09%的精度,96.87%,98.62%,分别,从而证明了FTVT在医学图像处理中的有效性和鲁棒性。
    Brain tumors occur due to the expansion of abnormal cell tissues and can be malignant (cancerous) or benign (not cancerous). Numerous factors such as the position, size, and progression rate are considered while detecting and diagnosing brain tumors. Detecting brain tumors in their initial phases is vital for diagnosis where MRI (magnetic resonance imaging) scans play an important role. Over the years, deep learning models have been extensively used for medical image processing. The current study primarily investigates the novel Fine-Tuned Vision Transformer models (FTVTs)-FTVT-b16, FTVT-b32, FTVT-l16, FTVT-l32-for brain tumor classification, while also comparing them with other established deep learning models such as ResNet50, MobileNet-V2, and EfficientNet - B0. A dataset with 7,023 images (MRI scans) categorized into four different classes, namely, glioma, meningioma, pituitary, and no tumor are used for classification. Further, the study presents a comparative analysis of these models including their accuracies and other evaluation metrics including recall, precision, and F1-score across each class. The deep learning models ResNet-50, EfficientNet-B0, and MobileNet-V2 obtained an accuracy of 96.5%, 95.1%, and 94.9%, respectively. Among all the FTVT models, FTVT-l16 model achieved a remarkable accuracy of 98.70% whereas other FTVT models FTVT-b16, FTVT-b32, and FTVT-132 achieved an accuracy of 98.09%, 96.87%, 98.62%, respectively, hence proving the efficacy and robustness of FTVT\'s in medical image processing.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    将自动分割方法结合到牙科X射线图像中,通过促进细致,完善了临床诊断和治疗计划的范例,牙齿结构和邻近组织的像素级关节。这是早期病理检测和细致的疾病进展监测的支柱。尽管如此,由于X射线成像的内在局限性,传统的分割框架经常会遇到重大挫折,包括受损的图像保真度,结构边界的模糊划定,以及牙髓等牙齿成分的复杂解剖结构,搪瓷,还有牙本质.为了克服这些障碍,我们提出了可变形卷积和Mamba集成网络,创新的2D牙科X射线图像分割架构,合并了一个合并结构可变形编码器,认知优化的语义增强模块,和分层收敛解码器。总的来说,这些组件支持多尺度全球功能的管理,加强特征表示的稳定性,并完善特征向量的合并。对14个基线的比较评估强调了其有效性,记录骰子系数增加了0.95%,第95个百分位数Hausdorff距离减少到7.494。
    The incorporation of automatic segmentation methodologies into dental X-ray images refined the paradigms of clinical diagnostics and therapeutic planning by facilitating meticulous, pixel-level articulation of both dental structures and proximate tissues. This underpins the pillars of early pathological detection and meticulous disease progression monitoring. Nonetheless, conventional segmentation frameworks often encounter significant setbacks attributable to the intrinsic limitations of X-ray imaging, including compromised image fidelity, obscured delineation of structural boundaries, and the intricate anatomical structures of dental constituents such as pulp, enamel, and dentin. To surmount these impediments, we propose the Deformable Convolution and Mamba Integration Network, an innovative 2D dental X-ray image segmentation architecture, which amalgamates a Coalescent Structural Deformable Encoder, a Cognitively-Optimized Semantic Enhance Module, and a Hierarchical Convergence Decoder. Collectively, these components bolster the management of multi-scale global features, fortify the stability of feature representation, and refine the amalgamation of feature vectors. A comparative assessment against 14 baselines underscores its efficacy, registering a 0.95% enhancement in the Dice Coefficient and a diminution of the 95th percentile Hausdorff Distance to 7.494.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    在受设备限制的临床条件下,实现轻量级的皮肤病变分割至关重要,因为它有助于将模型集成到各种医疗设备中,从而提高运营效率。然而,模型的轻量化设计可能面临精度下降,特别是当处理复杂的图像,如皮肤病变图像与不规则区域,模糊的边界,和超大的边界。为了应对这些挑战,我们提出了一个有效的轻量级注意网络(ELANet)用于皮肤病变分割任务。在ELANet,两种不同的注意机制的双边残差模块(BRM)可以实现信息互补,这增强了对空间和通道维度特征的敏感性,分别,然后将多个BRM堆叠起来,对输入信息进行有效的特征提取。此外,该网络通过多尺度注意力融合(MAF)操作放置不同尺度的特征图来获取全局信息并提高分割精度。最后,我们评估了ELANet在三个公开可用数据集上的性能,ISIC2016、ISIC2017和ISIC2018,实验结果表明,我们的算法可以达到89.87%,81.85%,三个参数为0.459M的数据集上的mIoU的82.87%,这是一个很好的平衡之间的准确性和亮度,是优于许多现有的分割方法。
    In clinical conditions limited by equipment, attaining lightweight skin lesion segmentation is pivotal as it facilitates the integration of the model into diverse medical devices, thereby enhancing operational efficiency. However, the lightweight design of the model may face accuracy degradation, especially when dealing with complex images such as skin lesion images with irregular regions, blurred boundaries, and oversized boundaries. To address these challenges, we propose an efficient lightweight attention network (ELANet) for the skin lesion segmentation task. In ELANet, two different attention mechanisms of the bilateral residual module (BRM) can achieve complementary information, which enhances the sensitivity to features in spatial and channel dimensions, respectively, and then multiple BRMs are stacked for efficient feature extraction of the input information. In addition, the network acquires global information and improves segmentation accuracy by putting feature maps of different scales through multi-scale attention fusion (MAF) operations. Finally, we evaluate the performance of ELANet on three publicly available datasets, ISIC2016, ISIC2017, and ISIC2018, and the experimental results show that our algorithm can achieve 89.87%, 81.85%, and 82.87% of the mIoU on the three datasets with a parametric of 0.459 M, which is an excellent balance between accuracy and lightness and is superior to many existing segmentation methods.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    骨小梁分析在了解骨骼健康和疾病中起着至关重要的作用。应用像骨质疏松症诊断。本文对三维骨小梁CT图像复原进行了全面的研究,解决这一领域的重大挑战。这项研究引入了一个骨干模型,级联-SwinUNETR,单视图三维CT图像复原。该模型利用具有监督和Swin-Transformer功能的深层聚合,在特征提取方面表现出色。此外,这项研究还带来了DVSR3D,双视图恢复模型,通过深度特征融合与注意力机制和自动编码器实现良好的性能。此外,介绍了一种无监督域自适应(UDA)方法,允许模型在没有额外标签的情况下适应输入数据分布,在现实世界的医疗应用中拥有巨大的潜力,并消除了对侵入性数据收集程序的需求。该研究还包括用于CT图像复原的新的双视图数据集的策展,解决Micro-CT中真实人体骨骼数据的稀缺性。最后,通过下游医学骨微结构测量验证了双视图方法。我们的贡献为骨小梁分析开辟了几条途径,有望改善骨健康评估和诊断的临床结果。
    Trabecular bone analysis plays a crucial role in understanding bone health and disease, with applications like osteoporosis diagnosis. This paper presents a comprehensive study on 3D trabecular computed tomography (CT) image restoration, addressing significant challenges in this domain. The research introduces a backbone model, Cascade-SwinUNETR, for single-view 3D CT image restoration. This model leverages deep layer aggregation with supervision and capabilities of Swin-Transformer to excel in feature extraction. Additionally, this study also brings DVSR3D, a dual-view restoration model, achieving good performance through deep feature fusion with attention mechanisms and Autoencoders. Furthermore, an Unsupervised Domain Adaptation (UDA) method is introduced, allowing models to adapt to input data distributions without additional labels, holding significant potential for real-world medical applications, and eliminating the need for invasive data collection procedures. The study also includes the curation of a new dual-view dataset for CT image restoration, addressing the scarcity of real human bone data in Micro-CT. Finally, the dual-view approach is validated through downstream medical bone microstructure measurements. Our contributions open several paths for trabecular bone analysis, promising improved clinical outcomes in bone health assessment and diagnosis.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    医学图像分割通常涉及多种组织类型和结构,包括血管分割和神经纤维束分割等任务。增强分割结果的连续性是医学图像分割的关键挑战,在临床应用需求的推动下,专注于疾病的定位和量化。在这项研究中,一种新颖的分割模型是专门为视网膜血管分割设计的,利用船只方位信息,边界约束,和连续性约束,以提高分割精度。为了实现这一点,我们将U-Net与长短期记忆网络(LSTM)级联。U-Net的特点是参数数量少,分割效率高,而LSTM提供参数共享功能。此外,我们引入了一个方向信息增强模块插入到模型的底层,通过方向卷积算子获得包含方向信息的特征图。此外,我们设计了一个新的混合损失函数,它由连接损失组成,边界损失,和交叉熵损失。实验结果表明,该模型在三个广泛认可的视网膜血管分割数据集上实现了出色的分割结果,CHASE_DB1,DRIVE,还有ARIA.
    Medical image segmentation commonly involves diverse tissue types and structures, including tasks such as blood vessel segmentation and nerve fiber bundle segmentation. Enhancing the continuity of segmentation outcomes represents a pivotal challenge in medical image segmentation, driven by the demands of clinical applications, focusing on disease localization and quantification. In this study, a novel segmentation model is specifically designed for retinal vessel segmentation, leveraging vessel orientation information, boundary constraints, and continuity constraints to improve segmentation accuracy. To achieve this, we cascade U-Net with a long-short-term memory network (LSTM). U-Net is characterized by a small number of parameters and high segmentation efficiency, while LSTM offers a parameter-sharing capability. Additionally, we introduce an orientation information enhancement module inserted into the model\'s bottom layer to obtain feature maps containing orientation information through an orientation convolution operator. Furthermore, we design a new hybrid loss function that consists of connectivity loss, boundary loss, and cross-entropy loss. Experimental results demonstrate that the model achieves excellent segmentation outcomes across three widely recognized retinal vessel segmentation datasets, CHASE_DB1, DRIVE, and ARIA.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    乳腺密度的评估,乳腺癌风险的关键指标,传统上由放射科医生通过乳房X线照相术图像的视觉检查来执行,利用乳腺成像报告和数据系统(BI-RADS)乳腺密度类别。然而,这种方法在观察者之间存在很大的可变性,导致密度评估和后续风险估计的不一致和潜在的不准确。为了解决这个问题,我们提出了一种基于深度学习的自动检测算法(DLAD),旨在自动评估乳腺密度。我们的多中心,多读者研究利用了来自三个机构的122个全视野数字乳房X线摄影研究的不同数据集(CC和MLO投影中的488张图像)。我们邀请了两位经验丰富的放射科医师进行回顾性分析,为72项乳房X线照相术研究(BI-RADSA类:18,BI-RADSB类:43,BI-RADSC类:7,BI-RADSD类:4)。然后将DLAD的功效与具有不同经验水平的五名独立放射科医师的表现进行比较。DLAD显示出强大的性能,达到0.819的准确度(95%CI:0.736-0.903),F1得分为0.798(0.594-0.905),精度为0.806(0.596-0.896),召回0.830(0.650-0.946),科恩的卡帕(κ)为0.708(0.562-0.841)。该算法实现了匹配的稳健性能,并且在四种情况下超过了单个放射科医生的稳健性能。统计分析并没有发现DLAD和放射科医师之间的准确性存在显着差异。强调该模型与专业放射科医生评估的竞争性诊断一致性。这些结果表明,基于深度学习的自动检测算法可以提高乳腺密度评估的准确性和一致性,为改善乳腺癌筛查结果提供了可靠的工具。
    The evaluation of mammographic breast density, a critical indicator of breast cancer risk, is traditionally performed by radiologists via visual inspection of mammography images, utilizing the Breast Imaging-Reporting and Data System (BI-RADS) breast density categories. However, this method is subject to substantial interobserver variability, leading to inconsistencies and potential inaccuracies in density assessment and subsequent risk estimations. To address this, we present a deep learning-based automatic detection algorithm (DLAD) designed for the automated evaluation of breast density. Our multicentric, multi-reader study leverages a diverse dataset of 122 full-field digital mammography studies (488 images in CC and MLO projections) sourced from three institutions. We invited two experienced radiologists to conduct a retrospective analysis, establishing a ground truth for 72 mammography studies (BI-RADS class A: 18, BI-RADS class B: 43, BI-RADS class C: 7, BI-RADS class D: 4). The efficacy of the DLAD was then compared to the performance of five independent radiologists with varying levels of experience. The DLAD showed robust performance, achieving an accuracy of 0.819 (95% CI: 0.736-0.903), along with an F1 score of 0.798 (0.594-0.905), precision of 0.806 (0.596-0.896), recall of 0.830 (0.650-0.946), and a Cohen\'s Kappa (κ) of 0.708 (0.562-0.841). The algorithm achieved robust performance that matches and in four cases exceeds that of individual radiologists. The statistical analysis did not reveal a significant difference in accuracy between DLAD and the radiologists, underscoring the model\'s competitive diagnostic alignment with professional radiologist assessments. These results demonstrate that the deep learning-based automatic detection algorithm can enhance the accuracy and consistency of breast density assessments, offering a reliable tool for improving breast cancer screening outcomes.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    肾脏肿瘤的发病率逐年递增。肾脏肿瘤的分割精度对诊断和治疗至关重要。
    为了提高准确性并减少人工参与,提出了一种基于深度学习的CT图像中肾脏和肾脏肿瘤自动分割方法。
    所提出的方法包括两个部分:对象检测和分割。我们首先使用一个模型来检测肾脏的位置,然后缩小分割范围,最后使用注意递归残差卷积网络进行分割。
    我们的模型在KiTS19数据集上获得了0.951的肾脏骰子得分和0.895的肿瘤骰子得分。实验结果表明,我们的模型显着提高了肾脏和肾脏肿瘤分割的准确性,并且优于其他高级方法。
    所提出的方法为在CT图像上准确分割肾脏和肾脏肿瘤提供了一种高效且自动的解决方案。此外,这项研究可以帮助放射科医师评估患者病情并做出明智的治疗决定.
    UNASSIGNED: The incidence of kidney tumors is progressively increasing each year. The precision of segmentation for kidney tumors is crucial for diagnosis and treatment.
    UNASSIGNED: To enhance accuracy and reduce manual involvement, propose a deep learning-based method for the automatic segmentation of kidneys and kidney tumors in CT images.
    UNASSIGNED: The proposed method comprises two parts: object detection and segmentation. We first use a model to detect the position of the kidney, then narrow the segmentation range, and finally use an attentional recurrent residual convolutional network for segmentation.
    UNASSIGNED: Our model achieved a kidney dice score of 0.951 and a tumor dice score of 0.895 on the KiTS19 dataset. Experimental results show that our model significantly improves the accuracy of kidney and kidney tumor segmentation and outperforms other advanced methods.
    UNASSIGNED: The proposed method provides an efficient and automatic solution for accurately segmenting kidneys and renal tumors on CT images. Additionally, this study can assist radiologists in assessing patients\' conditions and making informed treatment decisions.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

公众号