biomedical image processing

  • 文章类型: Journal Article
    口内环境的恶劣条件通常会导致低质量的照片和视频,阻碍进一步的临床诊断。要恢复这些数字记录,本研究提出了一种使用分段任意模型的实时交互式恢复系统。
    口内数字视频,通过口内摄像机从vident实验室数据集中获得,作为交互式恢复系统的输入。初始阶段采用利用段任何事物模型的交互式分段模块。随后,设计了实时帧内恢复模块和视频增强模块。系统地进行了一系列消融研究,以说明交互式恢复系统的优越设计。我们的定量评估标准包含修复质量,分割精度,和处理速度。此外,专家对处理后的视频的临床适用性进行了评估.
    广泛的实验证明了其在分割上的性能,平均交集为0.977。关于视频恢复,它导致可靠的性能,峰值信噪比为37.09,结构相似性指数度量为0.961。更多的可视化结果显示在https://yogurtsam上。github.io/iveproject页面。
    交互式修复系统展示了其为患者和牙医提供可靠且可控的口腔内视频修复的潜力。
    UNASSIGNED: Poor conditions in the intraoral environment often lead to low-quality photos and videos, hindering further clinical diagnosis. To restore these digital records, this study proposes a real-time interactive restoration system using segment anything model.
    UNASSIGNED: Intraoral digital videos, obtained from the vident-lab dataset through an intraoral camera, serve as the input for interactive restoration system. The initial phase employs an interactive segmentation module leveraging segment anything model. Subsequently, a real-time intraframe restoration module and a video enhancement module were designed. A series of ablation studies were systematically conducted to illustrate the superior design of interactive restoration system. Our quantitative evaluation criteria contain restoration quality, segmentation accuracy, and processing speed. Furthermore, the clinical applicability of the processed videos was evaluated by experts.
    UNASSIGNED: Extensive experiments demonstrated its performance on segmentation with a mean intersection-over-union of 0.977. On video restoration, it leads to reliable performances with peak signal-to-noise ratio of 37.09 and structural similarity index measure of 0.961, respectively. More visualization results are shown on the https://yogurtsam.github.io/iveproject page.
    UNASSIGNED: Interactive restoration system demonstrates its potential to serve patients and dentists with reliable and controllable intraoral video restoration.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    部分自动化机器人系统,如相机支架,代表了提高手术效率和精度的关键一步。因此,本文介绍了一种使用卷积神经网络在腹腔镜手术中实时工具定位的方法。提出的模型,基于两个串联的沙漏模块,可以同时定位两个手术工具。这项研究利用了三个数据集:ITAP数据集,除了两个公开可用的数据集,即AtlasDione和EndoVis挑战赛。提出了基于沙漏的模型的三种变体,使用最佳模型实现高精度(92.86%)和帧速率(27.64FPS),适合集成到机器人系统。对独立测试集的评估得出的准确性略低,表明泛化性有限。使用Grad-CAM技术进一步分析了该模型,以深入了解其功能。总的来说,这项工作为腹腔镜手术的自动化方面提出了一个有希望的解决方案,通过减少手动内窥镜操作的需要,有可能提高手术效率。
    Partially automated robotic systems, such as camera holders, represent a pivotal step towards enhancing efficiency and precision in surgical procedures. Therefore, this paper introduces an approach for real-time tool localization in laparoscopy surgery using convolutional neural networks. The proposed model, based on two Hourglass modules in series, can localize up to two surgical tools simultaneously. This study utilized three datasets: the ITAP dataset, alongside two publicly available datasets, namely Atlas Dione and EndoVis Challenge. Three variations of the Hourglass-based models were proposed, with the best model achieving high accuracy (92.86%) and frame rates (27.64 FPS), suitable for integration into robotic systems. An evaluation on an independent test set yielded slightly lower accuracy, indicating limited generalizability. The model was further analyzed using the Grad-CAM technique to gain insights into its functionality. Overall, this work presents a promising solution for automating aspects of laparoscopic surgery, potentially enhancing surgical efficiency by reducing the need for manual endoscope manipulation.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    甲皱毛细管镜检查是监测人体健康的重要手段。全景指甲折叠图像提高了检查的效率和准确性。然而,很少研究全景指甲折叠图像的获取,并且在对此类图像进行图像拼接时,存在匹配特征点很少的问题。因此,提出了一种基于血管轮廓增强的全景指甲图像拼接方法,首先通过对比度约束自适应直方图均衡化(CLAHE)对图像进行预处理,解决匹配特征点少的问题,双边滤波(BF),和锐化算法。然后使用快速鲁棒功能(SURF)成功拼接指甲褶皱血管的全景图像,快速近似最近邻库(FLANN)和随机样本协议(RANSAC)算法。实验结果表明,本文算法拼接的全景图像的视场宽度为7.43mm,提高了诊断的效率和准确性。
    Nail fold capillaroscopy is an important means of monitoring human health. Panoramic nail fold images improve the efficiency and accuracy of examinations. However, the acquisition of panoramic nail fold images is seldom studied and the problem manifests of few matching feature points when image stitching is used for such images. Therefore, this paper presents a method for panoramic nail fold image stitching based on vascular contour enhancement, which first solves the problem of few matching feature points by pre-processing the image with contrast-constrained adaptive histogram equalization (CLAHE), bilateral filtering (BF), and sharpening algorithms. The panoramic images of the nail fold blood vessels are then successfully stitched using the fast robust feature (SURF), fast library of approximate nearest neighbors (FLANN) and random sample agreement (RANSAC) algorithms. The experimental results show that the panoramic image stitched by this paper\'s algorithm has a field of view width of 7.43 mm, which improves the efficiency and accuracy of diagnosis.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    目的:虽然Yanagihara系统和House-Brackmann系统等主观方法是评估面瘫的标准,它们受到观察者内部和观察者之间可变性的限制。同时,诸如神经电图和肌电图等定量客观方法是耗时的。我们的目标是引入一个快速的,目标,和评价面部运动的定量方法。
    方法:我们开发了一种应用软件(app),利用iPhone的面部识别功能(AppleInc.,库比蒂诺,美国)用于面部运动评估。这个应用程序利用手机的前置摄像头,红外辐射,和红外摄像头,以提供详细的三维面部拓扑。它按区域定量比较左右面部运动,并显示患侧与相对侧的运动比率。使用该应用程序对正常和面部麻痹受试者进行评估,并与常规方法进行比较。
    结果:我们的应用程序提供了直观的用户体验,在一分钟内完成评估,因此证明了常规使用的实用性。其评估分数与柳原系统高度相关,House-Brackmann系统,和肌电图。此外,该应用程序在评估详细的面部运动方面优于传统方法。
    结论:我们新颖的iPhone应用程序为全面有效地评估面部麻痹提供了宝贵的工具。
    OBJECTIVE: While subjective methods like the Yanagihara system and the House-Brackmann system are standard in evaluating facial paralysis, they are limited by intra- and inter-observer variability. Meanwhile, quantitative objective methods such as electroneurography and electromyography are time-consuming. Our aim was to introduce a swift, objective, and quantitative method for evaluating facial movements.
    METHODS: We developed an application software (app) that utilizes the facial recognition functionality of the iPhone (Apple Inc., Cupertino, USA) for facial movement evaluation. This app leverages the phone\'s front camera, infrared radiation, and infrared camera to provide detailed three-dimensional facial topology. It quantitatively compares left and right facial movements by region and displays the movement ratio of the affected side to the opposite side. Evaluations using the app were conducted on both normal and facial palsy subjects and were compared with conventional methods.
    RESULTS: Our app provided an intuitive user experience, completing evaluations in under a minute, and thus proving practical for regular use. Its evaluation scores correlated highly with the Yanagihara system, the House-Brackmann system, and electromyography. Furthermore, the app outperformed conventional methods in assessing detailed facial movements.
    CONCLUSIONS: Our novel iPhone app offers a valuable tool for the comprehensive and efficient evaluation of facial palsy.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    目的:皮肤癌早期诊断可以挽救生命;然而,传统方法依赖于专家知识,耗时。这需要使用机器学习和深度学习的自动化系统。然而,现有的数据集通常集中在平坦的皮肤表面,忽略器官或附近病变的更复杂病例。
    方法:这项工作通过提出一种皮肤癌诊断方法来解决这一差距,该方法使用名为ASAN的数据集,该数据集涵盖了各种皮肤癌病例,但具有嘈杂的特征。为了克服嘈杂的功能问题,引入了一个名为SASAN的分割数据集,专注于基于感兴趣区域(ROI)提取的分类。这允许模型专注于图像内的关键区域,而忽略学习噪声特征。
    结果:各种深度学习分割模型,如UNet、LinkNet,PSPNet,和FPN在SASAN数据集上进行训练以执行基于分割的ROI提取。然后使用具有和不具有ROI提取的数据集进行分类。结果表明,ROI提取显着提高了这些模型在分类中的性能。这意味着SASAN在评估复杂皮肤癌病例的性能指标方面是有效的。
    结论:这项研究强调了扩展数据集以包括具有挑战性的场景和开发更好的分割方法以增强自动化皮肤癌诊断的重要性。SASAN数据集是旨在改进此类系统并最终有助于更好诊断结果的研究人员的宝贵工具。
    OBJECTIVE: Early skin cancer diagnosis can save lives; however, traditional methods rely on expert knowledge and can be time-consuming. This calls for automated systems using machine learning and deep learning. However, existing datasets often focus on flat skin surfaces, neglecting more complex cases on organs or with nearby lesions.
    METHODS: This work addresses this gap by proposing a skin cancer diagnosis methodology using a dataset named ASAN that covers diverse skin cancer cases but suffers from noisy features. To overcome the noisy feature problem, a segmentation dataset named SASAN is introduced, focusing on Region of Interest (ROI) extraction-based classification. This allows models to concentrate on critical areas within the images while ignoring learning the noisy features.
    RESULTS: Various deep learning segmentation models such as UNet, LinkNet, PSPNet, and FPN were trained on the SASAN dataset to perform segmentation-based ROI extraction. Classification was then performed using the dataset with and without ROI extraction. The results demonstrate that ROI extraction significantly improves the performance of these models in classification. This implies that SASAN is effective in evaluating performance metrics for complex skin cancer cases.
    CONCLUSIONS: This study highlights the importance of expanding datasets to include challenging scenarios and developing better segmentation methods to enhance automated skin cancer diagnosis. The SASAN dataset serves as a valuable tool for researchers aiming to improve such systems and ultimately contribute to better diagnostic outcomes.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    近年来,我们和其他人已经开发了非破坏性方法来获得临床活检和手术标本的三维(3D)病理学数据集。对于前列腺癌风险分层(预测),标准护理Gleason分级基于检查薄2D切片中前列腺的形态。这促使我们在我们的3D病理学数据集中执行前列腺的3D分割,以用于可以提供改进的预后性能的3D腺体特征的计算分析。
    为了促进前列腺癌风险评估,我们基于用苏木精和曙红(H&E)荧光类似物染色的人类前列腺活检的开放式光片显微镜数据集,开发了一种计算高效且准确的3D腺体分割深度学习模型。
    对于基于我们的H&E模拟3D病理学数据集的3D腺体分割,我们之前开发了一个基于混合深度学习和计算机视觉的管道,称为3D图像平移辅助分割(ITAS3D),这需要复杂的两阶段程序和繁琐的参数手动优化。为了简化这个程序,我们使用ITAS3D先前生成的3D腺体分割掩码作为直接基于端到端深度学习的分割模型的训练数据集,NNU-Net。该模型的输入是前列腺活检的3D病理学数据集,用廉价的H&E荧光模拟快速染色,输出是腺上皮的3D语义分割掩模,腺腔,和组织内周围的基质隔室。
    nnU-Net在3D腺体分割中表现出卓越的准确性,即使训练数据有限。此外,与以前的ITAS3D管道相比,nnU-Net操作更简单,更快捷,它可以保持良好的精度,即使在较低分辨率的输入。
    我们训练的基于DL的3D分割模型将促进未来的研究,以证明计算3D病理学在指导前列腺癌患者关键治疗决策方面的价值。
    UNASSIGNED: In recent years, we and others have developed non-destructive methods to obtain three-dimensional (3D) pathology datasets of clinical biopsies and surgical specimens. For prostate cancer risk stratification (prognostication), standard-of-care Gleason grading is based on examining the morphology of prostate glands in thin 2D sections. This motivates us to perform 3D segmentation of prostate glands in our 3D pathology datasets for the purposes of computational analysis of 3D glandular features that could offer improved prognostic performance.
    UNASSIGNED: To facilitate prostate cancer risk assessment, we developed a computationally efficient and accurate deep learning model for 3D gland segmentation based on open-top light-sheet microscopy datasets of human prostate biopsies stained with a fluorescent analog of hematoxylin and eosin (H&E).
    UNASSIGNED: For 3D gland segmentation based on our H&E-analog 3D pathology datasets, we previously developed a hybrid deep learning and computer vision-based pipeline, called image translation-assisted segmentation in 3D (ITAS3D), which required a complex two-stage procedure and tedious manual optimization of parameters. To simplify this procedure, we use the 3D gland-segmentation masks previously generated by ITAS3D as training datasets for a direct end-to-end deep learning-based segmentation model, nnU-Net. The inputs to this model are 3D pathology datasets of prostate biopsies rapidly stained with an inexpensive fluorescent analog of H&E and the outputs are 3D semantic segmentation masks of the gland epithelium, gland lumen, and surrounding stromal compartments within the tissue.
    UNASSIGNED: nnU-Net demonstrates remarkable accuracy in 3D gland segmentations even with limited training data. Moreover, compared with the previous ITAS3D pipeline, nnU-Net operation is simpler and faster, and it can maintain good accuracy even with lower-resolution inputs.
    UNASSIGNED: Our trained DL-based 3D segmentation model will facilitate future studies to demonstrate the value of computational 3D pathology for guiding critical treatment decisions for patients with prostate cancer.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    HT-29细胞系,来源于人类结肠癌,对生物学和癌症研究应用有价值。早期发现对于提高生存机会至关重要,研究人员正在引进新技术来准确诊断癌症。这项研究引入了一种高效的基于深度学习的方法来检测和计数结直肠癌细胞(HT-29)。结肠直肠癌细胞系是从一家公司获得的。Further,培养癌细胞,并且在实验室中进行了transwell实验,以通过荧光显微镜收集结直肠癌细胞图像的数据集。在566张照片中,80%被分配到训练组,其余20%被分配到测试集。通过集成YOLOv2、ResNet-50和ResNet-18架构来执行医学图像中的HT-29细胞检测和计数。ResNet-18的准确度为98.70%,ResNet-50的准确度为96.66%。该研究通过专注于检测和量化图像中的拥塞和重叠的结直肠癌细胞来实现其主要目标。这项创新工作构成了重叠癌细胞检测和计数的重大发展,为新的进步铺平道路,为研究和临床应用开辟新途径。研究人员可以通过探索ResNet和YOLO架构的变化来扩展研究,以优化对象检测性能。对实时部署策略的进一步研究将增强这些模型的实际适用性。
    The HT-29 cell line, derived from human colon cancer, is valuable for biological and cancer research applications. Early detection is crucial for improving the chances of survival, and researchers are introducing new techniques for accurate cancer diagnosis. This study introduces an efficient deep learning-based method for detecting and counting colorectal cancer cells (HT-29). The colorectal cancer cell line was procured from a company. Further, the cancer cells were cultured, and a transwell experiment was conducted in the lab to collect the dataset of colorectal cancer cell images via fluorescence microscopy. Of the 566 images, 80 % were allocated to the training set, and the remaining 20 % were assigned to the testing set. The HT-29 cell detection and counting in medical images is performed by integrating YOLOv2, ResNet-50, and ResNet-18 architectures. The accuracy achieved by ResNet-18 is 98.70 % and ResNet-50 is 96.66 %. The study achieves its primary objective by focusing on detecting and quantifying congested and overlapping colorectal cancer cells within the images. This innovative work constitutes a significant development in overlapping cancer cell detection and counting, paving the way for novel advancements and opening new avenues for research and clinical applications. Researchers can extend the study by exploring variations in ResNet and YOLO architectures to optimize object detection performance. Further investigation into real-time deployment strategies will enhance the practical applicability of these models.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    内窥镜检查有助于检查内脏,包括胃肠道.内窥镜装置由柔性管组成,相机和光源附接到柔性管。诊断过程在很大程度上取决于内窥镜图像的质量。这就是为什么内窥镜图像的视觉质量对患者护理有重大影响,医疗决策,和内镜治疗的效率。在这项研究中,提出了一种基于图像融合的内窥镜图像增强技术。我们的方法旨在通过首先从单个输入图像生成多个子图像来提高内窥镜图像的视觉质量,这些子图像在局部和全局对比度方面彼此互补。然后,每个子层都进行了新的小波变换和基于引导滤波器的分解技术。要生成最终的改进图像,最后利用适当的融合规则。在研究中对一组上消化道内窥镜图像进行了测试,以确认我们策略的有效性。定性和定量分析都表明,所提出的框架比一些最先进的算法性能更好。
    Endoscopies are helpful for examining internal organs, including the gastrointestinal tract. The endoscope device consists of a flexible tube to which a camera and light source are attached. The diagnostic process heavily depends on the quality of the endoscopic images. That is why the visual quality of endoscopic images has a significant effect on patient care, medical decision-making, and the efficiency of endoscopic treatments. In this study, we propose an endoscopic image enhancement technique based on image fusion. Our method aims to improve the visual quality of endoscopic images by first generating multiple sub images from the single input image which are complementary to one another in terms of local and global contrast. Then, each sub layer is subjected to a novel wavelet transform and guided filter-based decomposition technique. To generate the final improved image, appropriate fusion rules are utilized at the end. A set of upper gastrointestinal tract endoscopic images were put to the test in studies to confirm the efficacy of our strategy. Both qualitative and quantitative analyses show that the proposed framework performs better than some of the state-of-the-art algorithms.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    本文旨在探讨AI在提高使用X射线X射线照片识别髋部骨折的诊断准确性和效率方面的潜力。在研究中,我们训练了三种不同的深度学习模型,我们利用多数投票来评估他们的结果,旨在从X射线照片中获得最可靠和最精确的髋部骨折诊断。
    对从巴肯特大学附属的五家医院获得的10,849例AP骨盆X射线进行了初步研究。两名专业的整形外科医生最初将2,291张X射线照片标记为骨折,将8,558张X射线照片标记为非骨折。该算法在6,943(64%)的射线照片上进行了训练,在1,736(16%)射线照片上验证,并在2,170(20%)的射线照片上进行了测试,确保裂缝的均匀分布,年龄,和性别。我们采用了三种先进的深度学习架构,Xception(模型A),EfficientNet(模型B),和NFNet(型号C),通过多数投票技术(模型D)汇总最终决定。
    对于每个型号,我们获得了以下指标:对于模型A:F1评分0.895,准确性0.956,特异性0.973,敏感性0.893。对于模型B:F1评分0.900,准确度0.960,特异性0.991,灵敏度0.845。对于模型C:F1评分0.919,准确度0.966,特异性0.984,灵敏度0.899。对于模型D:F1评分0.929,准确度0.971,特异性0.991,灵敏度0.897。我们得出的结论是,就F1得分而言,模型D(多数投票)取得了最好的结果,准确度,和特异性值。
    我们的研究表明,通过投票汇总多个模型的决策所获得的结果,而不是仅仅依靠单一算法的决策,更加一致。这些算法的实际应用将是困难的,由于道德,legal,和保密问题,尽管取得了理论上的成功。开发成功的算法和方法不应被视为最终目标;重要的是要了解这些算法将如何在现实生活中使用。为了获得更一致的结果,从临床实践的反馈将是有益的。
    UNASSIGNED: This article was undertaken to explore the potential of AI in enhancing the diagnostic accuracy and efficiency in identifying hip fractures using X-ray radiographs. In the study, we trained three distinct deep learning models, and we utilized majority voting to evaluate their outcomes, aiming to yield the most reliable and precise diagnoses of hip fractures from X-ray radiographs.
    UNASSIGNED: An initial study was conducted of 10,849 AP pelvis X-rays obtained from five hospitals affiliated with Başkent University. Two expert orthopedic surgeons initially labeled 2,291 radiographs as fractures and 8,558 as non-fractures. The algorithm was trained on 6,943 (64%) radiographs, validated on 1,736 (16%) radiographs, and tested on 2,170 (20%) radiographs, ensuring an even distribution of fracture presence, age, and gender. We employed three advanced deep learning architectures, Xception (Model A), EfficientNet (Model B), and NfNet (Model C), with a final decision aggregated through a majority voting technique (Model D).
    UNASSIGNED: For each model, we achieved the following metrics:For Model A: F1 Score 0.895, Accuracy 0.956, Specificity 0.973, Sensitivity 0.893.For Model B: F1 Score 0.900, Accuracy 0.960, Specificity 0.991, Sensitivity 0.845.For Model C: F1 Score 0.919, Accuracy 0.966, Specificity 0.984, Sensitivity 0.899.For Model D: F1 Score 0.929, Accuracy 0.971, Specificity 0.991, Sensitivity 0.897.We concluded that Model D (majority voting) achieved the best results in terms of the F1 score, accuracy, and specificity values.
    UNASSIGNED: Our study demonstrates that the results obtained by aggregating the decisions of multiple models through voting, rather than relying solely on the decision of a single algorithm, are more consistent. The practical application of these algorithms will be difficult due to ethical, legal, and confidentiality issues, despite the theoretical success achieved. Developing successful algorithms and methodologies should not be viewed as the ultimate goal; it is important to understand how these algorithms will be used in real-life situations. In order to achieve more consistent results, feedback from clinical practice will be helpful.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    气管内插管(ETI)对于在紧急情况下确保气道至关重要。尽管人工智能算法经常用于分析医学图像,它们在基于紧急ETI期间捕获的图像评估口内结构方面的应用仍然有限.这项研究的目的是开发一种人工智能模型,用于使用视频喉镜(VL)图像分割口腔中的结构。
    来自54个VL视频,临床医生手动标记包括运动模糊的图像,雾天视觉,血,粘液,还有呕吐物.感兴趣的解剖结构包括舌头,会厌,声带,和角状软骨。EfficientNet-B5与DeepLabv3+,带U-Net的EffecientNet-B5,并使用配置的掩模R-卷积神经网络(CNN);EffecientNet-B5在ImageNet上进行预训练。使用骰子相似系数(DSC)来衡量模型的分割性能。准确性,召回,特异性,和F1评分用于根据地面实况和预测掩码之间的交叉值评估模型在瞄准结构方面的性能。
    舌头的DSC,会厌,声带,和用DeepLabv3+从EfficientNet-B5获得的角状软骨,带U-Net的EfficientNet-B5,和配置的MaskR-CNN模型分别为0.3351/0.7675/0.766/0.6539、0.0/0.7581/0.7395/0.6906和0.1167/0.7677/0.7207/0.57。此外,三个模型的处理速度(每秒帧数)分别为3、24和32。
    本研究中开发的算法可以帮助医疗提供者在紧急情况下进行ETI。
    UNASSIGNED: Endotracheal intubation (ETI) is critical to secure the airway in emergent situations. Although artificial intelligence algorithms are frequently used to analyze medical images, their application to evaluating intraoral structures based on images captured during emergent ETI remains limited. The aim of this study is to develop an artificial intelligence model for segmenting structures in the oral cavity using video laryngoscope (VL) images.
    UNASSIGNED: From 54 VL videos, clinicians manually labeled images that include motion blur, foggy vision, blood, mucus, and vomitus. Anatomical structures of interest included the tongue, epiglottis, vocal cord, and corniculate cartilage. EfficientNet-B5 with DeepLabv3+, EffecientNet-B5 with U-Net, and Configured Mask R-Convolution Neural Network (CNN) were used; EffecientNet-B5 was pretrained on ImageNet. Dice similarity coefficient (DSC) was used to measure the segmentation performance of the model. Accuracy, recall, specificity, and F1 score were used to evaluate the model\'s performance in targeting the structure from the value of the intersection over union between the ground truth and prediction mask.
    UNASSIGNED: The DSC of tongue, epiglottis, vocal cord, and corniculate cartilage obtained from the EfficientNet-B5 with DeepLabv3+, EfficientNet-B5 with U-Net, and Configured Mask R-CNN model were 0.3351/0.7675/0.766/0.6539, 0.0/0.7581/0.7395/0.6906, and 0.1167/0.7677/0.7207/0.57, respectively. Furthermore, the processing speeds (frames per second) of the three models stood at 3, 24, and 32, respectively.
    UNASSIGNED: The algorithm developed in this study can assist medical providers performing ETI in emergent situations.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

公众号