medical image

医学图像
  • 文章类型: Journal Article
    磁共振成像(MRI)在脑肿瘤分类中的应用受到传统诊断程序复杂、耗时的制约,主要是因为需要对几个地区进行全面评估。然而,深度学习(DL)的进步促进了自动化系统的开发,该系统可以改善医学图像的识别和评估,有效应对这些困难。卷积神经网络(CNN)已经成为图像分类和视觉感知的坚定工具。这项研究引入了一种创新的方法,将CNN与混合注意力机制相结合,对原发性脑肿瘤进行分类,包括神经胶质瘤,脑膜瘤,垂体,和无肿瘤病例。所提出的算法经过了来自文献中有据可查的基准数据的严格测试。它与建立的预训练模型如Xception、ResNet50V2、Densenet201、ResNet101V2和DenseNet169。该方法的性能指标显著,分类准确率为98.33%,准确率和召回率为98.30%,F1评分为98.20%。实验发现强调了新方法在识别最常见类型的脑肿瘤方面的优越性。此外,该方法表现出良好的泛化能力,使其成为医疗保健准确有效地诊断大脑状况的宝贵工具。
    The application of magnetic resonance imaging (MRI) in the classification of brain tumors is constrained by the complex and time-consuming characteristics of traditional diagnostics procedures, mainly because of the need for a thorough assessment across several regions. Nevertheless, advancements in deep learning (DL) have facilitated the development of an automated system that improves the identification and assessment of medical images, effectively addressing these difficulties. Convolutional neural networks (CNNs) have emerged as steadfast tools for image classification and visual perception. This study introduces an innovative approach that combines CNNs with a hybrid attention mechanism to classify primary brain tumors, including glioma, meningioma, pituitary, and no-tumor cases. The proposed algorithm was rigorously tested with benchmark data from well-documented sources in the literature. It was evaluated alongside established pre-trained models such as Xception, ResNet50V2, Densenet201, ResNet101V2, and DenseNet169. The performance metrics of the proposed method were remarkable, demonstrating classification accuracy of 98.33%, precision and recall of 98.30%, and F1-score of 98.20%. The experimental finding highlights the superior performance of the new approach in identifying the most frequent types of brain tumors. Furthermore, the method shows excellent generalization capabilities, making it an invaluable tool for healthcare in diagnosing brain conditions accurately and efficiently.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    在牙科领域,牙结石的存在是一个常见的问题。如果不及时解决,它有可能导致牙龈发炎和最终的牙齿脱落。Bitewing(BW)图像通过提供牙齿结构的全面视觉表示来发挥关键作用,允许牙医在临床评估期间精确检查难以到达的区域。这种视觉辅助明显有助于早期发现结石,促进及时干预并改善患者的总体预后。这项研究介绍了一种设计用于BW图像中牙结石检测的系统,利用YOLOv8的力量准确识别单个牙齿。该系统拥有令人印象深刻的97.48%的准确率,召回率(敏感度)为96.81%,特异性率为98.25%。此外,这项研究介绍了一种新的方法来增强齿间边缘通过先进的图像增强算法。该算法结合了中值滤波器和双边滤波器的使用,以改善卷积神经网络在对牙结石进行分类时的准确性。在图像增强之前,使用GoogLeNet实现的准确度为75.00%,显着提高到增强后的96.11%。这些结果具有简化牙科咨询的潜力,提高牙科服务的整体效率。
    In the field of dentistry, the presence of dental calculus is a commonly encountered issue. If not addressed promptly, it has the potential to lead to gum inflammation and eventual tooth loss. Bitewing (BW) images play a crucial role by providing a comprehensive visual representation of the tooth structure, allowing dentists to examine hard-to-reach areas with precision during clinical assessments. This visual aid significantly aids in the early detection of calculus, facilitating timely interventions and improving overall outcomes for patients. This study introduces a system designed for the detection of dental calculus in BW images, leveraging the power of YOLOv8 to identify individual teeth accurately. This system boasts an impressive precision rate of 97.48%, a recall (sensitivity) of 96.81%, and a specificity rate of 98.25%. Furthermore, this study introduces a novel approach to enhancing interdental edges through an advanced image-enhancement algorithm. This algorithm combines the use of a median filter and bilateral filter to refine the accuracy of convolutional neural networks in classifying dental calculus. Before image enhancement, the accuracy achieved using GoogLeNet stands at 75.00%, which significantly improves to 96.11% post-enhancement. These results hold the potential for streamlining dental consultations, enhancing the overall efficiency of dental services.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    软组织肉瘤,与宫颈癌和食道癌的发病率相似,来自各种软组织,如平滑肌,脂肪,和纤维组织。成像中肉瘤的有效分割对于准确诊断至关重要。
    本研究收集了45例大腿软组织肉瘤患者的多模态MRI图像,总计8,640张图像。这些图像由临床医生注释以描绘肉瘤区域,创建一个全面的数据集。我们基于UNet框架开发了一种新颖的细分模型,用残差网络和注意力机制增强,以改进特定于模态的信息提取。此外,采用自监督学习策略来优化编码器的特征提取能力。
    与单模态输入相比,新模型在使用多模态MRI图像时表现出优越的分割性能。通过各种实验设置验证了模型利用创建的数据集的有效性,确认增强的能力,以表征肿瘤区域在不同的模式。
    多模态MRI图像和先进的机器学习技术在我们的模型中的集成显着改善了大腿成像中软组织肉瘤的分割。这一进步有助于临床医生更好地诊断和了解患者的病情,利用不同成像方式的优势。进一步的研究可以探索这些技术在其他类型的软组织肉瘤和其他解剖部位的应用。
    UNASSIGNED: Soft tissue sarcomas, similar in incidence to cervical and esophageal cancers, arise from various soft tissues like smooth muscle, fat, and fibrous tissue. Effective segmentation of sarcomas in imaging is crucial for accurate diagnosis.
    UNASSIGNED: This study collected multi-modal MRI images from 45 patients with thigh soft tissue sarcoma, totaling 8,640 images. These images were annotated by clinicians to delineate the sarcoma regions, creating a comprehensive dataset. We developed a novel segmentation model based on the UNet framework, enhanced with residual networks and attention mechanisms for improved modality-specific information extraction. Additionally, self-supervised learning strategies were employed to optimize feature extraction capabilities of the encoders.
    UNASSIGNED: The new model demonstrated superior segmentation performance when using multi-modal MRI images compared to single-modal inputs. The effectiveness of the model in utilizing the created dataset was validated through various experimental setups, confirming the enhanced ability to characterize tumor regions across different modalities.
    UNASSIGNED: The integration of multi-modal MRI images and advanced machine learning techniques in our model significantly improves the segmentation of soft tissue sarcomas in thigh imaging. This advancement aids clinicians in better diagnosing and understanding the patient\'s condition, leveraging the strengths of different imaging modalities. Further studies could explore the application of these techniques to other types of soft tissue sarcomas and additional anatomical sites.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    背景:考虑到口腔肿瘤的稀有性和多样性,需要一个可靠的计算机辅助病理诊断系统。已成功设计出使用深度神经网络的基于内容的图像检索(CBIR)系统,用于数字病理学。由于缺乏针对口腔病理学量身定制的广泛的图像数据库和特征提取器,因此尚未研究用于口腔病理学的CBIR系统。
    方法:本研究使用从30类口腔肿瘤中构建的大型CBIR数据库来比较深度学习方法作为特征提取器。
    结果:通过使用自监督学习(SSL)方法(SimCLR为0.900,TiCo为0.897)在数据库图像上训练的模型,获得了接收器工作特征曲线(AUC)下的最高平均面积。使用来自使用智能手机拍摄的相同案例的查询图像验证了模型的可泛化性。当智能手机图像作为查询进行测试时,两种模型的平均AUC最高(SimCLR为0.871,TiCo为0.857)。我们通过评估前10个平均准确度并检查确切的诊断类别及其鉴别诊断类别来确保检索到的图像结果很容易观察到。
    结论:使用特定于目标部位的图像数据使用SSL方法训练深度学习模型有利于口腔肿瘤组织学中的CBIR任务,以获得组织学上有意义的结果和高性能。这一结果为CBIR系统的有效开发提供了见解,以帮助提高组织病理学诊断的准确性和速度,并在未来推进口腔肿瘤研究。
    BACKGROUND:  Oral tumors necessitate a dependable computer-assisted pathological diagnosis system considering their rarity and diversity. A content-based image retrieval (CBIR) system using deep neural networks has been successfully devised for digital pathology. No CBIR system for oral pathology has been investigated because of the lack of an extensive image database and feature extractors tailored to oral pathology.
    METHODS: This study uses a large CBIR database constructed from 30 categories of oral tumors to compare deep learning methods as feature extractors.
    RESULTS: The highest average area under the receiver operating characteristic curve (AUC) was achieved by models trained on database images using self-supervised learning (SSL) methods (0.900 with SimCLR and 0.897 with TiCo). The generalizability of the models was validated using query images from the same cases taken with smartphones. When smartphone images were tested as queries, both models yielded the highest mean AUC (0.871 with SimCLR and 0.857 with TiCo). We ensured the retrieved image result would be easily observed by evaluating the top 10 mean accuracies and checking for an exact diagnostic category and its differential diagnostic categories.
    CONCLUSIONS: Training deep learning models with SSL methods using image data specific to the target site is beneficial for CBIR tasks in oral tumor histology to obtain histologically meaningful results and high performance. This result provides insight into the effective development of a CBIR system to help improve the accuracy and speed of histopathology diagnosis and advance oral tumor research in the future.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    数字设备可以轻松伪造医学图像。医学图像中的复制移动伪造检测(CMFD)导致了无法访问高级医疗设备的地区的滥用。复制移动图像的伪造直接影响医生的决策。这里讨论的方法是检测医学图像伪造的最佳方法。
    所提出的方法基于一种进化算法,该算法可以很好地检测假块。在第一阶段,在离散余弦变换(DCT)的帮助下,图像被带到信号电平。然后通过应用离散小波变换(DWT)来准备分割。DWT的低低频段,具有最多的图像属性,被分成块。使用平衡优化算法搜索每个块。区块最有可能被选中,并生成最终图像。
    基于三个精度标准对所提出的方法进行了评估,召回,和F1,并获得90.07%,92.34%,和91.56%,分别。它优于在医学图像上研究的方法。
    得出的结论是,我们在医学图像中用于CMFD的方法更准确。
    UNASSIGNED: Digital devices can easily forge medical images. Copy-move forgery detection (CMFD) in medical image has led to abuses in areas where access to advanced medical devices is unavailable. Forgery of the copy-move image directly affects the doctor\'s decision. The method discussed here is an optimal method for detecting medical image forgery.
    UNASSIGNED: The proposed method is based on an evolutionary algorithm that can detect fake blocks well. In the first stage, the image is taken to the signal level with the help of a discrete cosine transform (DCT). It is then ready for segmentation by applying discrete wavelet transform (DWT). The low-low band of DWT, which has the most image properties, is divided into blocks. Each block is searched using the equilibrium optimization algorithm. The blocks are most likely to be selected, and the final image is generated.
    UNASSIGNED: The proposed method was evaluated based on three criteria of precision, recall, and F1 and obtained 90.07%, 92.34%, and 91.56%, respectively. It is superior to the methods studied on medical images.
    UNASSIGNED: It concluded that our method for CMFD in the medical images was more accurate.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    传统上,从医学图像中构建自动肌肉分割的训练数据集,涉及熟练的操作员,导致较高的劳动力成本和有限的可扩展性。为了解决这个问题,我们开发了一种工具,可以让非专家进行有效的注释,并评估其训练自动分割网络的有效性。我们的系统允许用户对模板三维(3D)解剖模型进行变形,以使用具有轴向独立控制点的自由变形来拟合目标磁共振图像。矢状,和日冕方向。这种方法通过允许非专家直观地调整模型来简化注释过程,启用模板中所有肌肉的同时注释。我们评估了由非专家执行的工具辅助分割的质量,与专家分割相比,Dice系数大于0.75,没有明显的错误,例如错误标记相邻肌肉或省略肌肉组织。使用此工具创建的数据集训练的自动分割网络显示出与使用专家生成的数据集训练的网络相当或优于的性能。这种创新的工具大大减少了与自动肌肉分割数据集创建相关的时间和劳动力成本,潜在的革命性的医学图像标注和加速在各种临床应用中基于深度学习的分割网络的发展。
    Traditionally, constructing training datasets for automatic muscle segmentation from medical images involved skilled operators, leading to high labor costs and limited scalability. To address this issue, we developed a tool that enables efficient annotation by non-experts and assessed its effectiveness for training an automatic segmentation network. Our system allows users to deform a template three-dimensional (3D) anatomical model to fit a target magnetic-resonance image using free-form deformation with independent control points for axial, sagittal, and coronal directions. This method simplifies the annotation process by allowing non-experts to intuitively adjust the model, enabling simultaneous annotation of all muscles in the template. We evaluated the quality of the tool-assisted segmentation performed by non-experts, which achieved a Dice coefficient greater than 0.75 compared to expert segmentation, without significant errors such as mislabeling adjacent muscles or omitting musculature. An automatic segmentation network trained with datasets created using this tool demonstrated performance comparable to or superior to that of networks trained with expert-generated datasets. This innovative tool significantly reduces the time and labor costs associated with dataset creation for automatic muscle segmentation, potentially revolutionizing medical image annotation and accelerating the development of deep learning-based segmentation networks in various clinical applications.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: English Abstract
    Objective:To build a VGG-based computer-aided diagnostic model for chronic sinusitis and evaluate its efficacy. Methods:①A total of 5 000 frames of diagnosed sinus CT images were collected. The normal group consisted of 1 000 frames(250 frames each of maxillary sinus, frontal sinus, septal sinus, and pterygoid sinus), while the abnormal group consisted of 4 000 frames(1 000 frames each of maxillary sinusitis, frontal sinusitis, septal sinusitis, and pterygoid sinusitis). ②The models were trained and simulated to obtain five classification models for the normal group, the pteroid sinusitis group, the frontal sinusitis group, the septal sinusitis group and the maxillary sinusitis group, respectively. The classification efficacy of the models was evaluated objectively in six dimensions: accuracy, precision, sensitivity, specificity, interpretation time and area under the ROC curve(AUC). ③Two hundred randomly selected images were read by the model with three groups of physicians(low, middle and high seniority) to constitute a comparative experiment. The efficacy of the model was objectively evaluated using the aforementioned evaluation indexes in conjunction with clinical analysis. Results:①Simulation experiment: The overall recognition accuracy of the model is 83.94%, with a precision of 89.52%, sensitivity of 83.94%, specificity of 95.99%, and the average interpretation time of each frame is 0.2 s. The AUC for sphenoid sinusitis was 0.865(95%CI 0.849-0.881), for frontal sinusitis was 0.924(0.991-0.936), for ethmoidoid sinusitis was 0.895(0.880-0.909), and for maxillary sinusitis was 0.974(0.967-0.982). ②Comparison experiment: In terms of recognition accuracy, the model was 84.52%, while the low-seniority physicians group was 78.50%, the middle-seniority physicians group was 80.50%, and the seniority physicians group was 83.50%; In terms of recognition accuracy, the model was 85.67%, the low seniority physicians group was 79.72%, the middle seniority physicians group was 82.67%, and the high seniority physicians group was 83.66%. In terms of recognition sensitivity, the model was 84.52%, the low seniority group was 78.50%, the middle seniority group was 80.50%, and the high seniority group was 83.50%. In terms of recognition specificity, the model was 96.58%, the low-seniority physicians group was 94.63%, the middle-seniority physicians group was 95.13%, and the seniority physicians group was 95.88%. In terms of time consumption, the average image per frame of the model is 0.20 s, the average image per frame of the low-seniority physicians group is 2.35 s, the average image per frame of the middle-seniority physicians group is 1.98 s, and the average image per frame of the senior physicians group is 2.19 s. Conclusion:This study demonstrates the potential of a deep learning-based artificial intelligence diagnostic model for chronic sinusitis to classify and diagnose chronic sinusitis; the deep learning-based artificial intelligence diagnosis model for chronic sinusitis has good classification performance and high diagnostic efficacy.
    目的:搭建基于VGG的慢性鼻窦炎计算机辅助诊断模型,并评价其效能。 方法:①收集5 000帧已确诊的鼻窦CT图像,将其分为正常组1 000帧图像(其中,正常的上颌窦、额窦、筛窦、蝶窦影像图像各250帧)及异常组4 000帧图像(其中,上颌窦炎、额窦炎、筛窦炎、蝶窦炎影像图像各1 000帧),对图像进行大小归一化及分割预处理;②训练模型并对其进行仿真实验,分别得到正常组,蝶窦炎组,额窦炎组,筛窦炎组以及上颌窦炎组5个分类模型,从准确度、精确度、灵敏度、特异度、判读时间及ROC曲线下面积(AUC)6个维度,客观评价模型的分类效能;③随机选取200帧图像,通过模型与低年资医师组、中年资医师组、高年资医师组分别阅片构成对比试验,结合临床通过以上评价指标客观评价模型的效能。 结果:①仿真实验:整个模型的识别准确度为83.94%,精确度为89.52%,灵敏度为83.94%,特异度为95.99%,平均每帧图像判读时间为0.20 s;蝶窦炎的AUC为0.865(95%CI 0.849~0.881),额窦炎的AUC为0.924(0.911~0.936),筛窦炎的AUC为0.895(0.880~0.909),上颌窦炎的AUC为0.974(0.967~0.982)。②对比实验:在识别准确度上,模型为84.52%,低年资医师组为78.5%、中年资医师组为80.5%,高年资医师组为83.5%;在识别精确度上,模型为85.67%,低年资医师组为79.72%,中年资医师组为82.67%,高年资医师组为83.66%;在识别灵敏度上,模型为84.52%,低年资医师组为78.50%,中年资医师组为80.50%,高年资医师组为83.50%;在识别特异度上,模型为96.58%,低年资医师组为94.63%,中年资医师组为95.13%,高年资医师组为95.88%;在耗时上,模型平均每帧图像为0.20 s,低年资医师组平均每帧图像为2.35 s,中年资医师组平均每帧图像为1.98 s,高年资医师组平均每帧图像为2.19 s。 结论:本研究强调了基于深度学习的慢性鼻窦炎人工智能诊断模型分类诊断慢性鼻窦炎的可能性;基于深度学习的慢性鼻窦炎人工智能诊断模型分类性能好,具有较高的诊断效能。.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    背景:机器学习(ML)模型可以产生更快,更准确的医疗诊断;但是,开发ML模型受到缺乏高质量标记训练数据的限制。众包标签是一种潜在的解决方案,但可能会受到对标签质量的担忧的限制。
    目的:本研究旨在研究具有持续绩效评估的游戏化众包平台,用户反馈,基于绩效的激励措施可以在医学影像数据上产生专家质量标签。
    方法:在这项诊断比较研究中,回顾性收集了203例急诊科患者的2384例肺超声夹。共有6位肺部超声专家将这些夹子中的393个归类为没有B线,一条或多条离散的B线,或融合的B线创建2套参考标准数据集(195个训练剪辑和198个测试剪辑)。集合分别用于(1)在游戏化的众包平台上训练用户,以及(2)将所得人群标签的一致性与各个专家与参考标准的一致性进行比较。人群意见来自DiagnosUs(Centaur实验室)iOS应用程序用户超过8天,根据过去的性能进行过滤,使用多数规则聚合,并分析了与专家标记的夹子的固定测试集相比的标签一致性。主要结果是将经过整理的人群意见的标签一致性与训练有素的专家比较,以对肺部超声夹子上的B线进行分类。
    结果:我们的临床数据集包括平均年龄为60.0(SD19.0)岁的患者;105例(51.7%)患者为女性,114例(56.1%)患者为白人。在195个训练剪辑中,专家共识标签分布为114(58%)无B线,56(29%)离散B线,和25(13%)融合的B系。在198个测试夹上,专家共识标签分布为138(70%)无B线,36条(18%)离散B线,和24(12%)融合的B系。总的来说,收集了426个独特用户的99,238条意见。在198个夹子的测试集上,个别专家相对于参考标准的平均标签一致性为85.0%(SE2.0),与87.9%的众包标签一致性相比(P=0.15)。当个别专家的意见与参考标准标签进行比较时,多数投票创建的不包括他们自己的意见,人群一致性高于个别专家对参考标准的平均一致性(87.4%vs80.8%,SE1.6表示专家一致性;P<.001)。具有离散B线的剪辑在人群共识和专家共识中的分歧最大。使用随机抽样的人群意见子集,7种经过质量过滤的意见足以达到接近最大的人群一致性。
    结论:通过游戏化方法对肺部超声夹进行B线分类的众包标签达到了专家级的准确性。这表明游戏化众包在有效生成用于训练ML系统的标记图像数据集方面具有战略作用。
    BACKGROUND: Machine learning (ML) models can yield faster and more accurate medical diagnoses; however, developing ML models is limited by a lack of high-quality labeled training data. Crowdsourced labeling is a potential solution but can be constrained by concerns about label quality.
    OBJECTIVE: This study aims to examine whether a gamified crowdsourcing platform with continuous performance assessment, user feedback, and performance-based incentives could produce expert-quality labels on medical imaging data.
    METHODS: In this diagnostic comparison study, 2384 lung ultrasound clips were retrospectively collected from 203 emergency department patients. A total of 6 lung ultrasound experts classified 393 of these clips as having no B-lines, one or more discrete B-lines, or confluent B-lines to create 2 sets of reference standard data sets (195 training clips and 198 test clips). Sets were respectively used to (1) train users on a gamified crowdsourcing platform and (2) compare the concordance of the resulting crowd labels to the concordance of individual experts to reference standards. Crowd opinions were sourced from DiagnosUs (Centaur Labs) iOS app users over 8 days, filtered based on past performance, aggregated using majority rule, and analyzed for label concordance compared with a hold-out test set of expert-labeled clips. The primary outcome was comparing the labeling concordance of collated crowd opinions to trained experts in classifying B-lines on lung ultrasound clips.
    RESULTS: Our clinical data set included patients with a mean age of 60.0 (SD 19.0) years; 105 (51.7%) patients were female and 114 (56.1%) patients were White. Over the 195 training clips, the expert-consensus label distribution was 114 (58%) no B-lines, 56 (29%) discrete B-lines, and 25 (13%) confluent B-lines. Over the 198 test clips, expert-consensus label distribution was 138 (70%) no B-lines, 36 (18%) discrete B-lines, and 24 (12%) confluent B-lines. In total, 99,238 opinions were collected from 426 unique users. On a test set of 198 clips, the mean labeling concordance of individual experts relative to the reference standard was 85.0% (SE 2.0), compared with 87.9% crowdsourced label concordance (P=.15). When individual experts\' opinions were compared with reference standard labels created by majority vote excluding their own opinion, crowd concordance was higher than the mean concordance of individual experts to reference standards (87.4% vs 80.8%, SE 1.6 for expert concordance; P<.001). Clips with discrete B-lines had the most disagreement from both the crowd consensus and individual experts with the expert consensus. Using randomly sampled subsets of crowd opinions, 7 quality-filtered opinions were sufficient to achieve near the maximum crowd concordance.
    CONCLUSIONS: Crowdsourced labels for B-line classification on lung ultrasound clips via a gamified approach achieved expert-level accuracy. This suggests a strategic role for gamified crowdsourcing in efficiently generating labeled image data sets for training ML systems.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    目前,脑肿瘤是非常有害和普遍的。深度学习技术,包括CNN,UNet,变压器,在脑肿瘤分割中应用多年,取得了一定的成功。然而,传统的CNN和UNet捕获的全球信息不足,和变压器不能提供足够的本地信息。将来自Transformer的全局信息与卷积的局部信息融合是改善脑肿瘤分割的重要一步。我们提出了群体归一化洗牌和增强型信道自注意网络(GETNet),将纯变压器结构与基于VT-UNet的卷积运算相结合的网络,它考虑了全球和本地信息。该网络包括所提出的组归一化混洗块(GNS)和增强型信道自注意块(ECSA)。在VT编码器块之后和下采样块之前使用GNS以改进信息提取。将ECSA模块添加到瓶颈层,以有效地利用底层中的详细特征的特性。我们还对BraTS2021数据集进行了实验,以证明我们网络的性能。Dice系数(Dice)评分结果表明,整个肿瘤(WT)区域的值,肿瘤核心(TC),和增强肿瘤(ET)分别为91.77、86.03和83.64。结果表明,与十一个以上的基准测试相比,该模型实现了最先进的性能。
    Currently, brain tumors are extremely harmful and prevalent. Deep learning technologies, including CNNs, UNet, and Transformer, have been applied in brain tumor segmentation for many years and have achieved some success. However, traditional CNNs and UNet capture insufficient global information, and Transformer cannot provide sufficient local information. Fusing the global information from Transformer with the local information of convolutions is an important step toward improving brain tumor segmentation. We propose the Group Normalization Shuffle and Enhanced Channel Self-Attention Network (GETNet), a network combining the pure Transformer structure with convolution operations based on VT-UNet, which considers both global and local information. The network includes the proposed group normalization shuffle block (GNS) and enhanced channel self-attention block (ECSA). The GNS is used after the VT Encoder Block and before the downsampling block to improve information extraction. An ECSA module is added to the bottleneck layer to utilize the characteristics of the detailed features in the bottom layer effectively. We also conducted experiments on the BraTS2021 dataset to demonstrate the performance of our network. The Dice coefficient (Dice) score results show that the values for the regions of the whole tumor (WT), tumor core (TC), and enhancing tumor (ET) were 91.77, 86.03, and 83.64, respectively. The results show that the proposed model achieves state-of-the-art performance compared with more than eleven benchmarks.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    心脏计算机断层扫描(CT)和磁共振成像(MRI)的自动分割在心血管疾病的预防和治疗中起着至关重要的作用。在这项研究中,我们提出了一种基于多尺度的高效网络,多头自我注意(MSMHSA)机制。这种机制的结合使我们能够实现更大的感受野,有助于在CT和MRI图像中准确分割整个心脏结构。在这个网络中,从浅层特征提取网络中提取的特征经过MHSA机制,与人类视觉密切相关,使得上下文语义信息的提取更加全面和准确。为了提高不同尺寸的心脏子结构分割的精度,我们提出的方法在不同的尺度上引入了三个MHSA网络。这种方法允许通过调整分割图像的大小来微调微目标分割的准确性。我们方法的有效性在多模式全心脏分割(MM-WHS)挑战2017数据集上得到了严格验证,在心脏CT和MRI图像中展示有竞争力的结果和七个心脏亚结构的准确分割。通过与先进的基于变压器的模型的对比实验,我们的研究提供了令人信服的证据,尽管基于变压器的模型取得了显著成就,CNN模型和自我注意力的融合仍然是双模态全心脏分割的一种简单而高效的方法.
    The automatic segmentation of cardiac computed tomography (CT) and magnetic resonance imaging (MRI) plays a pivotal role in the prevention and treatment of cardiovascular diseases. In this study, we propose an efficient network based on the multi-scale, multi-head self-attention (MSMHSA) mechanism. The incorporation of this mechanism enables us to achieve larger receptive fields, facilitating the accurate segmentation of whole heart structures in both CT and MRI images. Within this network, features extracted from the shallow feature extraction network undergo a MHSA mechanism that closely aligns with human vision, resulting in the extraction of contextual semantic information more comprehensively and accurately. To improve the precision of cardiac substructure segmentation across varying sizes, our proposed method introduces three MHSA networks at distinct scales. This approach allows for fine-tuning the accuracy of micro-object segmentation by adapting the size of the segmented images. The efficacy of our method is rigorously validated on the Multi-Modality Whole Heart Segmentation (MM-WHS) Challenge 2017 dataset, demonstrating competitive results and the accurate segmentation of seven cardiac substructures in both cardiac CT and MRI images. Through comparative experiments with advanced transformer-based models, our study provides compelling evidence that despite the remarkable achievements of transformer-based models, the fusion of CNN models and self-attention remains a simple yet highly effective approach for dual-modality whole heart segmentation.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

公众号