medical image

医学图像
  • 文章类型: Journal Article
    Colorectal cancer (CRC) is a common malignant tumor that seriously threatens human health. CRC presents a formidable challenge in terms of accurate identification due to its indistinct boundaries. With the widespread adoption of convolutional neural networks (CNNs) in image processing, leveraging CNNs for automatic classification and segmentation holds immense potential for enhancing the efficiency of colorectal cancer recognition and reducing treatment costs. This paper explores the imperative necessity for applying CNNs in clinical diagnosis of CRC. It provides an elaborate overview on research advancements pertaining to CNNs and their improved models in CRC classification and segmentation. Furthermore, this work summarizes the ideas and common methods for optimizing network performance and discusses the challenges faced by CNNs as well as future development trends in their application towards CRC classification and segmentation, thereby promoting their utilization within clinical diagnosis.
    结直肠癌是一种常见的胃肠道恶性肿瘤,严重威胁人类健康。由于结直肠癌区边界模糊,使得对结直肠癌的准确识别存在很大挑战。随着卷积神经网络在图像处理领域应用的普及,利用卷积神经网络进行结直肠癌的自动分类与分割,在提高结直肠癌识别效率、降低癌症治疗成本方面具有很大潜力。本文论述了卷积神经网络在结直肠癌临床诊断中应用的必要性;详细介绍了目前卷积神经网络及其改进型在结直肠癌分类和分割两个部分中的研究进展;总结了对于网络性能优化的思路和常用方法,并讨论了卷积神经网络应用在结直肠癌分类与分割中所面对的挑战和未来的发展趋势,以促进卷积神经网络在结直肠癌临床诊断中的应用。.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    本文介绍了有效的医学图像目标分割任何事物模型(EMedSAM),解决了使用SAM进行医学图像分割任务的高计算需求和有限的适应性。我们提出了一个小说,紧凑型图像编码器,DD-TinyViT,旨在通过一种称为med-adapter的创新参数调整方法来提高分割效率。轻量级DD-TinyViT编码器是使用解耦蒸馏方法从众所周知的ViT-H导出的。MEdSAM对特定结构的分割和识别能力通过med-adapter提高,动态调整专门用于医学成像的模型参数。我们使用公共FLARE2022数据集和浙江大学医学院附属第一医院的数据集对EMedSAM进行了广泛的测试。结果表明,我们的模型在多器官和肺分割任务中都优于现有的最新模型。
    This paper introduces the efficient medical-images-aimed segment anything model (EMedSAM), addressing the high computational demands and limited adaptability of using SAM for medical image segmentation tasks. We present a novel, compact image encoder, DD-TinyViT, designed to enhance segmentation efficiency through an innovative parameter tuning method called med-adapter. The lightweight DD-TinyViT encoder is derived from the well-known ViT-H using a decoupled distillation approach.The segmentation and recognition capabilities of EMedSAM for specific structures are improved by med-adapter, which dynamically adjusts the model parameters specifically for medical imaging. We conducted extensive testing on EMedSAM using the public FLARE 2022 dataset and datasets from the First Hospital of Zhejiang University School of Medicine. The results demonstrate that our model outperforms existing state-of-the-art models in both multi-organ and lung segmentation tasks.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    在临床实践中,肺静脉的解剖分类在房颤射频消融手术的术前评估中起着至关重要的作用。肺静脉解剖结构的准确分类有助于医生选择合适的标测电极,避免引起肺动脉高压。由于肺静脉的解剖分类多种多样,以及数据分布的不平衡,深度学习模型在提取深度特征时往往表现出较差的表达能力,导致误判,影响分类精度。因此,为了解决左心房肺静脉分类不平衡的问题,本文提出了一种融合多尺度特征增强注意力和双特征提取分类器的网络,叫做DECNet。多尺度特征增强注意力利用多尺度信息引导深层特征的强化,生成通道权重和空间权重,增强深层特征的表达能力。双特征提取分类器为每个类别分配固定数量的通道,平等地评估所有类别,从而缓解了数据失衡导致的学习偏差和过拟合。通过将两者结合起来,增强了深层特征的表达能力,实现对左心房肺静脉形态的准确分类,为后续临床治疗提供支持。所提出的方法是在辽宁省人民医院提供的数据集和公开的DermaMNIST数据集上进行评估的,平均准确率为78.81%和83.44%,分别,证明了所提出方法的有效性。
    In clinical practice, the anatomical classification of pulmonary veins plays a crucial role in the preoperative assessment of atrial fibrillation radiofrequency ablation surgery. Accurate classification of pulmonary vein anatomy assists physicians in selecting appropriate mapping electrodes and avoids causing pulmonary arterial hypertension. Due to the diverse and subtly different anatomical classifications of pulmonary veins, as well as the imbalance in data distribution, deep learning models often exhibit poor expression capability in extracting deep features, leading to misjudgments and affecting classification accuracy. Therefore, in order to solve the problem of unbalanced classification of left atrial pulmonary veins, this paper proposes a network integrating multi-scale feature-enhanced attention and dual-feature extraction classifiers, called DECNet. The multi-scale feature-enhanced attention utilizes multi-scale information to guide the reinforcement of deep features, generating channel weights and spatial weights to enhance the expression capability of deep features. The dual-feature extraction classifier assigns a fixed number of channels to each category, equally evaluating all categories, thus alleviating the learning bias and overfitting caused by data imbalance. By combining the two, the expression capability of deep features is strengthened, achieving accurate classification of left atrial pulmonary vein morphology and providing support for subsequent clinical treatment. The proposed method is evaluated on datasets provided by the People\'s Hospital of Liaoning Province and the publicly available DermaMNIST dataset, achieving average accuracies of 78.81% and 83.44%, respectively, demonstrating the effectiveness of the proposed approach.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    磁共振成像(MRI)在脑肿瘤分类中的应用受到传统诊断程序复杂、耗时的制约,主要是因为需要对几个地区进行全面评估。然而,深度学习(DL)的进步促进了自动化系统的开发,该系统可以改善医学图像的识别和评估,有效应对这些困难。卷积神经网络(CNN)已经成为图像分类和视觉感知的坚定工具。这项研究引入了一种创新的方法,将CNN与混合注意力机制相结合,对原发性脑肿瘤进行分类,包括神经胶质瘤,脑膜瘤,垂体,和无肿瘤病例。所提出的算法经过了来自文献中有据可查的基准数据的严格测试。它与建立的预训练模型如Xception、ResNet50V2、Densenet201、ResNet101V2和DenseNet169。该方法的性能指标显著,分类准确率为98.33%,准确率和召回率为98.30%,F1评分为98.20%。实验发现强调了新方法在识别最常见类型的脑肿瘤方面的优越性。此外,该方法表现出良好的泛化能力,使其成为医疗保健准确有效地诊断大脑状况的宝贵工具。
    The application of magnetic resonance imaging (MRI) in the classification of brain tumors is constrained by the complex and time-consuming characteristics of traditional diagnostics procedures, mainly because of the need for a thorough assessment across several regions. Nevertheless, advancements in deep learning (DL) have facilitated the development of an automated system that improves the identification and assessment of medical images, effectively addressing these difficulties. Convolutional neural networks (CNNs) have emerged as steadfast tools for image classification and visual perception. This study introduces an innovative approach that combines CNNs with a hybrid attention mechanism to classify primary brain tumors, including glioma, meningioma, pituitary, and no-tumor cases. The proposed algorithm was rigorously tested with benchmark data from well-documented sources in the literature. It was evaluated alongside established pre-trained models such as Xception, ResNet50V2, Densenet201, ResNet101V2, and DenseNet169. The performance metrics of the proposed method were remarkable, demonstrating classification accuracy of 98.33%, precision and recall of 98.30%, and F1-score of 98.20%. The experimental finding highlights the superior performance of the new approach in identifying the most frequent types of brain tumors. Furthermore, the method shows excellent generalization capabilities, making it an invaluable tool for healthcare in diagnosing brain conditions accurately and efficiently.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    软组织肉瘤,与宫颈癌和食道癌的发病率相似,来自各种软组织,如平滑肌,脂肪,和纤维组织。成像中肉瘤的有效分割对于准确诊断至关重要。
    本研究收集了45例大腿软组织肉瘤患者的多模态MRI图像,总计8,640张图像。这些图像由临床医生注释以描绘肉瘤区域,创建一个全面的数据集。我们基于UNet框架开发了一种新颖的细分模型,用残差网络和注意力机制增强,以改进特定于模态的信息提取。此外,采用自监督学习策略来优化编码器的特征提取能力。
    与单模态输入相比,新模型在使用多模态MRI图像时表现出优越的分割性能。通过各种实验设置验证了模型利用创建的数据集的有效性,确认增强的能力,以表征肿瘤区域在不同的模式。
    多模态MRI图像和先进的机器学习技术在我们的模型中的集成显着改善了大腿成像中软组织肉瘤的分割。这一进步有助于临床医生更好地诊断和了解患者的病情,利用不同成像方式的优势。进一步的研究可以探索这些技术在其他类型的软组织肉瘤和其他解剖部位的应用。
    UNASSIGNED: Soft tissue sarcomas, similar in incidence to cervical and esophageal cancers, arise from various soft tissues like smooth muscle, fat, and fibrous tissue. Effective segmentation of sarcomas in imaging is crucial for accurate diagnosis.
    UNASSIGNED: This study collected multi-modal MRI images from 45 patients with thigh soft tissue sarcoma, totaling 8,640 images. These images were annotated by clinicians to delineate the sarcoma regions, creating a comprehensive dataset. We developed a novel segmentation model based on the UNet framework, enhanced with residual networks and attention mechanisms for improved modality-specific information extraction. Additionally, self-supervised learning strategies were employed to optimize feature extraction capabilities of the encoders.
    UNASSIGNED: The new model demonstrated superior segmentation performance when using multi-modal MRI images compared to single-modal inputs. The effectiveness of the model in utilizing the created dataset was validated through various experimental setups, confirming the enhanced ability to characterize tumor regions across different modalities.
    UNASSIGNED: The integration of multi-modal MRI images and advanced machine learning techniques in our model significantly improves the segmentation of soft tissue sarcomas in thigh imaging. This advancement aids clinicians in better diagnosing and understanding the patient\'s condition, leveraging the strengths of different imaging modalities. Further studies could explore the application of these techniques to other types of soft tissue sarcomas and additional anatomical sites.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: English Abstract
    Objective:To build a VGG-based computer-aided diagnostic model for chronic sinusitis and evaluate its efficacy. Methods:①A total of 5 000 frames of diagnosed sinus CT images were collected. The normal group consisted of 1 000 frames(250 frames each of maxillary sinus, frontal sinus, septal sinus, and pterygoid sinus), while the abnormal group consisted of 4 000 frames(1 000 frames each of maxillary sinusitis, frontal sinusitis, septal sinusitis, and pterygoid sinusitis). ②The models were trained and simulated to obtain five classification models for the normal group, the pteroid sinusitis group, the frontal sinusitis group, the septal sinusitis group and the maxillary sinusitis group, respectively. The classification efficacy of the models was evaluated objectively in six dimensions: accuracy, precision, sensitivity, specificity, interpretation time and area under the ROC curve(AUC). ③Two hundred randomly selected images were read by the model with three groups of physicians(low, middle and high seniority) to constitute a comparative experiment. The efficacy of the model was objectively evaluated using the aforementioned evaluation indexes in conjunction with clinical analysis. Results:①Simulation experiment: The overall recognition accuracy of the model is 83.94%, with a precision of 89.52%, sensitivity of 83.94%, specificity of 95.99%, and the average interpretation time of each frame is 0.2 s. The AUC for sphenoid sinusitis was 0.865(95%CI 0.849-0.881), for frontal sinusitis was 0.924(0.991-0.936), for ethmoidoid sinusitis was 0.895(0.880-0.909), and for maxillary sinusitis was 0.974(0.967-0.982). ②Comparison experiment: In terms of recognition accuracy, the model was 84.52%, while the low-seniority physicians group was 78.50%, the middle-seniority physicians group was 80.50%, and the seniority physicians group was 83.50%; In terms of recognition accuracy, the model was 85.67%, the low seniority physicians group was 79.72%, the middle seniority physicians group was 82.67%, and the high seniority physicians group was 83.66%. In terms of recognition sensitivity, the model was 84.52%, the low seniority group was 78.50%, the middle seniority group was 80.50%, and the high seniority group was 83.50%. In terms of recognition specificity, the model was 96.58%, the low-seniority physicians group was 94.63%, the middle-seniority physicians group was 95.13%, and the seniority physicians group was 95.88%. In terms of time consumption, the average image per frame of the model is 0.20 s, the average image per frame of the low-seniority physicians group is 2.35 s, the average image per frame of the middle-seniority physicians group is 1.98 s, and the average image per frame of the senior physicians group is 2.19 s. Conclusion:This study demonstrates the potential of a deep learning-based artificial intelligence diagnostic model for chronic sinusitis to classify and diagnose chronic sinusitis; the deep learning-based artificial intelligence diagnosis model for chronic sinusitis has good classification performance and high diagnostic efficacy.
    目的:搭建基于VGG的慢性鼻窦炎计算机辅助诊断模型,并评价其效能。 方法:①收集5 000帧已确诊的鼻窦CT图像,将其分为正常组1 000帧图像(其中,正常的上颌窦、额窦、筛窦、蝶窦影像图像各250帧)及异常组4 000帧图像(其中,上颌窦炎、额窦炎、筛窦炎、蝶窦炎影像图像各1 000帧),对图像进行大小归一化及分割预处理;②训练模型并对其进行仿真实验,分别得到正常组,蝶窦炎组,额窦炎组,筛窦炎组以及上颌窦炎组5个分类模型,从准确度、精确度、灵敏度、特异度、判读时间及ROC曲线下面积(AUC)6个维度,客观评价模型的分类效能;③随机选取200帧图像,通过模型与低年资医师组、中年资医师组、高年资医师组分别阅片构成对比试验,结合临床通过以上评价指标客观评价模型的效能。 结果:①仿真实验:整个模型的识别准确度为83.94%,精确度为89.52%,灵敏度为83.94%,特异度为95.99%,平均每帧图像判读时间为0.20 s;蝶窦炎的AUC为0.865(95%CI 0.849~0.881),额窦炎的AUC为0.924(0.911~0.936),筛窦炎的AUC为0.895(0.880~0.909),上颌窦炎的AUC为0.974(0.967~0.982)。②对比实验:在识别准确度上,模型为84.52%,低年资医师组为78.5%、中年资医师组为80.5%,高年资医师组为83.5%;在识别精确度上,模型为85.67%,低年资医师组为79.72%,中年资医师组为82.67%,高年资医师组为83.66%;在识别灵敏度上,模型为84.52%,低年资医师组为78.50%,中年资医师组为80.50%,高年资医师组为83.50%;在识别特异度上,模型为96.58%,低年资医师组为94.63%,中年资医师组为95.13%,高年资医师组为95.88%;在耗时上,模型平均每帧图像为0.20 s,低年资医师组平均每帧图像为2.35 s,中年资医师组平均每帧图像为1.98 s,高年资医师组平均每帧图像为2.19 s。 结论:本研究强调了基于深度学习的慢性鼻窦炎人工智能诊断模型分类诊断慢性鼻窦炎的可能性;基于深度学习的慢性鼻窦炎人工智能诊断模型分类性能好,具有较高的诊断效能。.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    目前,脑肿瘤是非常有害和普遍的。深度学习技术,包括CNN,UNet,变压器,在脑肿瘤分割中应用多年,取得了一定的成功。然而,传统的CNN和UNet捕获的全球信息不足,和变压器不能提供足够的本地信息。将来自Transformer的全局信息与卷积的局部信息融合是改善脑肿瘤分割的重要一步。我们提出了群体归一化洗牌和增强型信道自注意网络(GETNet),将纯变压器结构与基于VT-UNet的卷积运算相结合的网络,它考虑了全球和本地信息。该网络包括所提出的组归一化混洗块(GNS)和增强型信道自注意块(ECSA)。在VT编码器块之后和下采样块之前使用GNS以改进信息提取。将ECSA模块添加到瓶颈层,以有效地利用底层中的详细特征的特性。我们还对BraTS2021数据集进行了实验,以证明我们网络的性能。Dice系数(Dice)评分结果表明,整个肿瘤(WT)区域的值,肿瘤核心(TC),和增强肿瘤(ET)分别为91.77、86.03和83.64。结果表明,与十一个以上的基准测试相比,该模型实现了最先进的性能。
    Currently, brain tumors are extremely harmful and prevalent. Deep learning technologies, including CNNs, UNet, and Transformer, have been applied in brain tumor segmentation for many years and have achieved some success. However, traditional CNNs and UNet capture insufficient global information, and Transformer cannot provide sufficient local information. Fusing the global information from Transformer with the local information of convolutions is an important step toward improving brain tumor segmentation. We propose the Group Normalization Shuffle and Enhanced Channel Self-Attention Network (GETNet), a network combining the pure Transformer structure with convolution operations based on VT-UNet, which considers both global and local information. The network includes the proposed group normalization shuffle block (GNS) and enhanced channel self-attention block (ECSA). The GNS is used after the VT Encoder Block and before the downsampling block to improve information extraction. An ECSA module is added to the bottleneck layer to utilize the characteristics of the detailed features in the bottom layer effectively. We also conducted experiments on the BraTS2021 dataset to demonstrate the performance of our network. The Dice coefficient (Dice) score results show that the values for the regions of the whole tumor (WT), tumor core (TC), and enhancing tumor (ET) were 91.77, 86.03, and 83.64, respectively. The results show that the proposed model achieves state-of-the-art performance compared with more than eleven benchmarks.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    心脏计算机断层扫描(CT)和磁共振成像(MRI)的自动分割在心血管疾病的预防和治疗中起着至关重要的作用。在这项研究中,我们提出了一种基于多尺度的高效网络,多头自我注意(MSMHSA)机制。这种机制的结合使我们能够实现更大的感受野,有助于在CT和MRI图像中准确分割整个心脏结构。在这个网络中,从浅层特征提取网络中提取的特征经过MHSA机制,与人类视觉密切相关,使得上下文语义信息的提取更加全面和准确。为了提高不同尺寸的心脏子结构分割的精度,我们提出的方法在不同的尺度上引入了三个MHSA网络。这种方法允许通过调整分割图像的大小来微调微目标分割的准确性。我们方法的有效性在多模式全心脏分割(MM-WHS)挑战2017数据集上得到了严格验证,在心脏CT和MRI图像中展示有竞争力的结果和七个心脏亚结构的准确分割。通过与先进的基于变压器的模型的对比实验,我们的研究提供了令人信服的证据,尽管基于变压器的模型取得了显著成就,CNN模型和自我注意力的融合仍然是双模态全心脏分割的一种简单而高效的方法.
    The automatic segmentation of cardiac computed tomography (CT) and magnetic resonance imaging (MRI) plays a pivotal role in the prevention and treatment of cardiovascular diseases. In this study, we propose an efficient network based on the multi-scale, multi-head self-attention (MSMHSA) mechanism. The incorporation of this mechanism enables us to achieve larger receptive fields, facilitating the accurate segmentation of whole heart structures in both CT and MRI images. Within this network, features extracted from the shallow feature extraction network undergo a MHSA mechanism that closely aligns with human vision, resulting in the extraction of contextual semantic information more comprehensively and accurately. To improve the precision of cardiac substructure segmentation across varying sizes, our proposed method introduces three MHSA networks at distinct scales. This approach allows for fine-tuning the accuracy of micro-object segmentation by adapting the size of the segmented images. The efficacy of our method is rigorously validated on the Multi-Modality Whole Heart Segmentation (MM-WHS) Challenge 2017 dataset, demonstrating competitive results and the accurate segmentation of seven cardiac substructures in both cardiac CT and MRI images. Through comparative experiments with advanced transformer-based models, our study provides compelling evidence that despite the remarkable achievements of transformer-based models, the fusion of CNN models and self-attention remains a simple yet highly effective approach for dual-modality whole heart segmentation.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    医学影像学是临床诊断的重要工具。然而,医生准备影像诊断报告非常耗时且容易出错。因此,有必要开发一些自动生成医学影像报告的方法。目前,医学成像报告生成的任务至少在两个方面具有挑战性:(1)医学图像彼此非常相似。正常和异常图像之间以及不同异常图像之间的差异通常是微不足道的;(2)在生成的报告中描述异常发现的不相关或不正确的关键词导致错误通信。在本文中,我们提出了一个由四个模块组成的医学图像报告生成框架,包括一个变压器编码器,MIX-MLP多标签分类网络,基于协同注意机制(CAM)的语义和视觉特征融合,和分层LSTM解码器。Transformer编码器可用于学习图像和标签之间的远程依赖关系,有效地提取图像的视觉和语义特征,并在视觉和语义信息之间建立长期的依赖关系,以准确地从图像中提取异常特征。MIX-MLP多标签分类网络,共同关注机制和分层LSTM网络可以更好地识别异常,实现视觉和文本对齐融合和多标签诊断分类,更好地促进报告生成。在两个广泛使用的放射学报告数据集上进行的实验结果,IUX射线和MIMIC-CXR,表明我们提出的框架在自然语言生成指标和临床疗效评估指标方面优于当前的报告生成模型。这项工作的代码可在https://github.com/watersunhznu/LIFMRG在线获得。
    Medical imaging is an important tool for clinical diagnosis. Nevertheless, it is very time-consuming and error-prone for physicians to prepare imaging diagnosis reports. Therefore, it is necessary to develop some methods to generate medical imaging reports automatically. Currently, the task of medical imaging report generation is challenging in at least two aspects: (1) medical images are very similar to each other. The differences between normal and abnormal images and between different abnormal images are usually trivial; (2) unrelated or incorrect keywords describing abnormal findings in the generated reports lead to mis-communications. In this paper, we propose a medical image report generation framework composed of four modules, including a Transformer encoder, a MIX-MLP multi-label classification network, a co-attention mechanism (CAM) based semantic and visual feature fusion, and a hierarchical LSTM decoder. The Transformer encoder can be used to learn long-range dependencies between images and labels, effectively extract visual and semantic features of images, and establish long-term dependent relationships between visual and semantic information to accurately extract abnormal features from images. The MIX-MLP multi-label classification network, the co-attention mechanism and the hierarchical LSTM network can better identify abnormalities, achieving visual and text alignment fusion and multi-label diagnostic classification to better facilitate report generation. The results of the experiments performed on two widely used radiology report datasets, IU X-RAY and MIMIC-CXR, show that our proposed framework outperforms current report generation models in terms of both natural linguistic generation metrics and clinical efficacy assessment metrics. The code of this work is available online at https://github.com/watersunhznu/LIFMRG.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    医学成像是当前癌症诊断的重要工具。然而,医学图像的质量通常会受到影响,以最大程度地降低与患者图像采集相关的潜在风险。计算机辅助诊断系统近年来取得了重大进展。这些系统利用计算机算法来识别医学图像中的异常特征,协助放射科医生提高诊断准确性,并实现图像和疾病解释的一致性。重要的是,医学图像的质量,作为目标数据,通过人工智能算法确定可实现的性能水平。然而,医学图像的像素值范围不同于通常通过人工智能算法处理的数字图像的像素值范围,并且盲目地结合这些数据进行训练可能会导致算法性能欠佳。在这项研究中,我们提出了一种集成通用数字图像处理和医学图像处理模块的医学图像增强方案。该方案旨在通过赋予医学图像高对比度和平滑特性来增强医学图像数据。我们进行了实验测试,以证明该方案在提高医学图像分割算法性能方面的有效性。
    Medical imaging serves as a crucial tool in current cancer diagnosis. However, the quality of medical images is often compromised to minimize the potential risks associated with patient image acquisition. Computer-aided diagnosis systems have made significant advancements in recent years. These systems utilize computer algorithms to identify abnormal features in medical images, assisting radiologists in improving diagnostic accuracy and achieving consistency in image and disease interpretation. Importantly, the quality of medical images, as the target data, determines the achievable level of performance by artificial intelligence algorithms. However, the pixel value range of medical images differs from that of the digital images typically processed via artificial intelligence algorithms, and blindly incorporating such data for training can result in suboptimal algorithm performance. In this study, we propose a medical image-enhancement scheme that integrates generic digital image processing and medical image processing modules. This scheme aims to enhance medical image data by endowing them with high-contrast and smooth characteristics. We conducted experimental testing to demonstrate the effectiveness of this scheme in improving the performance of a medical image segmentation algorithm.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

公众号