deep-learning

深度学习
  • 文章类型: Journal Article
    血管周围间隙(EPVS)在老年人中很常见,但它们的神经病理学相关性尚不清楚,主要是因为迄今为止的大多数工作都依赖于视觉评定量表和/或临床队列.本研究首先开发了一种用于自动分割的深度学习模型,离体脑MRI中EPVS的定位和定量,然后用这个模型来研究神经病理学,在817名接受尸检的社区老年人中,EPVS的临床和认知相关性.新方法在检测小至3mm3的EPVS时具有很高的灵敏度,良好的分割精度和一致性。大多数EPVS位于额叶,但密度最高的是在基底神经节。大脑中,特别是额叶中的EPVS与梗死相关,与其他神经病理无关。而颞叶和枕叶EPVS与脑淀粉样血管病有关。大多数脑叶的EPVS也与糖尿病相关,与神经病理学无关。而基底神经节EPVS与高血压独立相关,支持从糖尿病和高血压到EPVS的独立途径的概念。最后,EPVS与较低的认知表现相关,独立于神经病理学和临床变量,提示EPVS代表导致认知降低的额外异常。
    Enlarged perivascular spaces (EPVS) are common in older adults, but their neuropathologic correlates are unclear mainly because most work to date has relied on visual rating scales and/or clinical cohorts. The present study first developed a deep-learning model for automatic segmentation, localization and quantification of EPVS in ex vivo brain MRI, and then used this model to investigate the neuropathologic, clinical and cognitive correlates of EPVS in 817 community-based older adults that underwent autopsy. The new method exhibited high sensitivity in detecting EPVS as small as 3 mm3, good segmentation accuracy and consistency. Most EPVS were located in the frontal lobe, but the highest density was observed in the basal ganglia. EPVS in the cerebrum and specifically in the frontal lobe were associated with infarcts independent of other neuropathologies, while temporal and occipital EPVS were associated with cerebral amyloid angiopathy. EPVS in most brain lobes were also associated with diabetes mellitus independently of neuropathologies, while basal ganglia EPVS were independently associated with hypertension, supporting the notion of independent pathways from diabetes and hypertension to EPVS. Finally, EPVS were associated with lower cognitive performance independently of neuropathologies and clinical variables, suggesting that EPVS represent additional abnormalities contributing to lower cognition.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    最近的全球健康危机,COVID-19是一场严重的全球健康危机,深刻影响了人们的生活方式。使用医学图像从类似的胸部异常中检测此类疾病是一项具有挑战性的任务。因此,在临床治疗中,端到端自动化系统的要求是非常必要的。这样,这项工作提出了一种基于挤压和激发注意力的ResNet50(SEA-ResNet50)模型,用于利用胸部X射线数据检测COVID-19。这里,这个想法在于使用挤压和激励注意力机制改进ResNet50的剩余单位。为了进一步增强,Ranger优化器和自适应Mish激活函数用于改进SEA-ResNet50模型的特征学习。为了评估,利用了两个公开的COVID-19射线照相数据集。在实验期间,胸部X射线输入图像被增强,以针对四个输出类别进行稳健评估,即正常,肺炎,肺混浊,和COVID-19。然后对SEA-ResNet50模型与VGG-16,Xception,ResNet18、ResNet50和DenseNet121体系结构。与现有的CNN架构相比,所提出的SEA-ResNet50框架以及Ranger优化器和自适应Mish激活提供了98.38%(多类)和99.29%(二元分类)的最大分类精度。所提出的方法比其他方法获得了0.975(多分类)和0.98(二元分类)的最高Kappa验证分数。此外,使用可解释人工智能(XAI)模型来表示异常区域的显著性图的可视化,从而提高疾病诊断的可解释性。
    A recent global health crisis, COVID-19 is a significant global health crisis that has profoundly affected lifestyles. The detection of such diseases from similar thoracic anomalies using medical images is a challenging task. Thus, the requirement of an end-to-end automated system is vastly necessary in clinical treatments. In this way, the work proposes a Squeeze-and-Excitation Attention-based ResNet50 (SEA-ResNet50) model for detecting COVID-19 utilizing chest X-ray data. Here, the idea lies in improving the residual units of ResNet50 using the squeeze-and-excitation attention mechanism. For further enhancement, the Ranger optimizer and adaptive Mish activation function are employed to improve the feature learning of the SEA-ResNet50 model. For evaluation, two publicly available COVID-19 radiographic datasets are utilized. The chest X-ray input images are augmented during experimentation for robust evaluation against four output classes namely normal, pneumonia, lung opacity, and COVID-19. Then a comparative study is done for the SEA-ResNet50 model against VGG-16, Xception, ResNet18, ResNet50, and DenseNet121 architectures. The proposed framework of SEA-ResNet50 together with the Ranger optimizer and adaptive Mish activation provided maximum classification accuracies of 98.38% (multiclass) and 99.29% (binary classification) as compared with the existing CNN architectures. The proposed method achieved the highest Kappa validation scores of 0.975 (multiclass) and 0.98 (binary classification) over others. Furthermore, the visualization of the saliency maps of the abnormal regions is represented using the explainable artificial intelligence (XAI) model, thereby enhancing interpretability in disease diagnosis.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    在单个机构临床应用中,使用两种不同的商用基于深度学习的自动分割(DLAS)工具,评估计算机断层扫描图像的头颈部区域中的危险器官(OAR)自动分割。
    根据已发布的40例临床头颈部癌(HNC)病例的计算机断层扫描(pCT)图像规划指南,临床医生对22例OAR进行了手动轮廓绘制。使用两个基于深度学习的自动分割模型ManteiaAccuContour和MIMProtégéAI为每位患者生成自动轮廓。然后使用Sørensen-Dice相似性系数(DSC)和平均距离(MD)指标将自动轮廓(AC)的准确性和完整性与专家轮廓(EC)进行比较。
    使用AccuContour生成22个OAR和使用ProtégéAI生成17个OAR的AC,平均轮廓生成时间分别为1分钟/患者和5分钟/患者。下颌骨的EC和AC一致性最高(DSC0.90±0.16)和(DSC0.91±0.03),AccuContour和ProtégéAI的chiasm(DSC0.28±0.14)和(DSC0.30±0.14)分别最低。使用AccuContour,22个OAR轮廓中有10个的平均MD<1mm,6OAR为1-2mm,6OAR为2-3mm。对于ProtégéAI,17个OAR中有8个的平均距离<1mm,6OAR为1-2mm,3OAR为2-3mm。
    两种DLAS程序都被证明是有价值的工具,可以显着减少在头颈部区域生成大量OAR轮廓所需的时间,即使在实施治疗计划之前可能需要手动编辑AC。获得的DSC和MD与评估各种其他DLAS解决方案的其他研究中报道的相类似。尽管如此,CT图像中具有非理想对比度的小体积结构,比如神经,非常具有挑战性,需要额外的解决方案才能取得足够的成果。
    UNASSIGNED: To evaluate organ at risk (OAR) auto-segmentation in the head and neck region of computed tomography images using two different commercially available deep-learning-based auto-segmentation (DLAS) tools in a single institutional clinical applications.
    UNASSIGNED: Twenty-two OARs were manually contoured by clinicians according to published guidelines on planning computed tomography (pCT) images for 40 clinical head and neck cancer (HNC) cases. Automatic contours were generated for each patient using two deep-learning-based auto-segmentation models-Manteia AccuContour and MIM ProtégéAI. The accuracy and integrity of autocontours (ACs) were then compared to expert contours (ECs) using the Sørensen-Dice similarity coefficient (DSC) and Mean Distance (MD) metrics.
    UNASSIGNED: ACs were generated for 22 OARs using AccuContour and 17 OARs using ProtégéAI with average contour generation time of 1 min/patient and 5 min/patient respectively. EC and AC agreement was highest for the mandible (DSC 0.90 ± 0.16) and (DSC 0.91 ± 0.03), and lowest for the chiasm (DSC 0.28 ± 0.14) and (DSC 0.30 ± 0.14) for AccuContour and ProtégéAI respectively. Using AccuContour, the average MD was<1mm for 10 of the 22 OARs contoured, 1-2mm for 6 OARs, and 2-3mm for 6 OARs. For ProtégéAI, the average mean distance was<1mm for 8 out of 17 OARs, 1-2mm for 6 OARs, and 2-3mm for 3 OARs.
    UNASSIGNED: Both DLAS programs were proven to be valuable tools to significantly reduce the time required to generate large amounts of OAR contours in the head and neck region, even though manual editing of ACs is likely needed prior to implementation into treatment planning. The DSCs and MDs achieved were similar to those reported in other studies that evaluated various other DLAS solutions. Still, small volume structures with nonideal contrast in CT images, such as nerves, are very challenging and will require additional solutions to achieve sufficient results.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    在放射干预期间监测辐射剂量和时间参数至关重要,尤其是在神经介入手术中,如用栓塞线圈治疗动脉瘤。提出的算法检测医学图像中这些栓塞线圈的存在。它建立了一个边界框作为自动准直的参考,主要目标是通过主动优化图像质量同时最小化患者剂量来提高神经介入手术的效率和安全性。
    在我们的研究中评估了两种不同的方法。第一个涉及深度学习,采用具有ResNet-50FPN作为骨干的FasterR-CNN模型和RetinaNet模型。第二种方法利用经典的斑点检测方法,作为比较的基准。
    我们进行了五次交叉验证,我们表现最好的模型在验证数据的所有折叠上实现了0.84的平均mAP@75,在独立测试数据上实现了0.94的平均mAP@75。由于我们使用放大的边界框,没有必要实现地面实况和预测之间的100%重叠。为了突出我们算法的实际应用,我们进行了一个由合金丝构成的线圈的模拟,有效地展示了自动准直的实施。这导致剂量面积乘积显著减少,通过尽量减少散射辐射,这意味着患者和医务人员的随机风险降低。此外,我们的算法有助于在窄准直期间避免X射线血管造影图像中的极端亮度或黑暗,最终简化了医生的准直过程。
    据我们所知,这标志着对成功检测栓塞线圈的方法的初步尝试,展示了将检测结果集成到X射线血管造影系统中的扩展应用。我们提出的方法具有更广泛的应用潜力,允许其扩展以检测介入程序中使用的其他医疗对象。
    UNASSIGNED: Monitoring radiation dose and time parameters during radiological interventions is crucial, especially in neurointerventional procedures, such as aneurysm treatment with embolization coils. The algorithm presented detects the presence of these embolization coils in medical images. It establishes a bounding box as a reference for automated collimation, with the primary objective being to enhance the efficiency and safety of neurointerventional procedures by actively optimizing image quality while minimizing patient dose.
    UNASSIGNED: Two distinct methodologies are evaluated in our study. The first involves deep learning, employing the Faster R-CNN model with a ResNet-50 FPN as a backbone and a RetinaNet model. The second method utilizes a classical blob detection approach, serving as a benchmark for comparison.
    UNASSIGNED: We performed a fivefold cross-validation, and our top-performing model achieved mean mAP@75 of 0.84 across all folds on validation data and mean mAP@75 of 0.94 on independent test data. Since we use an upscaled bounding box, achieving 100% overlap between ground truth and prediction is not necessary. To highlight the real-world applications of our algorithm, we conducted a simulation featuring a coil constructed from an alloy wire, effectively showcasing the implementation of automatic collimation. This resulted in a notable reduction in the dose area product, signifying the reduction of stochastic risks for both patients and medical staff by minimizing scatter radiation. Additionally, our algorithm assists in avoiding extreme brightness or darkness in X-ray angiography images during narrow collimation, ultimately streamlining the collimation process for physicians.
    UNASSIGNED: To our knowledge, this marks the initial attempt at an approach successfully detecting embolization coils, showcasing the extended applications of integrating detection results into the X-ray angiography system. The method we present has the potential for broader application, allowing its extension to detect other medical objects utilized in interventional procedures.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    包装椰子水提供各种选择,从纯净到添加糖和其他添加剂的那些。虽然椰子水的纯度因其健康益处而受到尊重,它的受欢迎程度也使其面临潜在的掺假和虚假陈述。为了解决这一问题,我们的研究结合傅里叶变换红外光谱(FTIR)和机器学习技术,通过分类模型检测椰子水中潜在的掺假物。该数据集包含来自椰子水样品的红外光谱,其中掺有15种不同类型的潜在糖替代品,包括:糖,人造甜味剂,和糖醇。红外光与分子键的相互作用产生独特的分子指纹,形成我们分析的基础。与先前的研究不同,该研究主要依赖于基于线性的化学计量学来检测掺假物,我们的研究探索了线性,非线性,和组合特征提取模型。通过使用主成分分析(PCA)和t分布随机邻居嵌入(t-SNE)开发交互式应用程序,通过增强的可视化和模式识别,简化了非靶向糖掺假物检测。使用集成学习随机森林(RF)和深度学习一维卷积神经网络(1DCNN)进行有针对性的分析,实现了更高的分类精度(95%和96%,分别)与相同数据集上77%的稀疏偏最小二乘判别分析(SPLS-DA)和88%的支持向量机(SVM)相比。CNN展示的分类准确性通过其对原始数据进行训练和测试的能力得到了卓越效率的补充。
    Packaged coconut water offers various options, from pure to those with added sugars and other additives. While the purity of coconut water is esteemed for its health benefits, its popularity also exposes it to potential adulteration and misrepresentation. To address this concern, our study combines Fourier transform infrared spectroscopy (FTIR) and machine learning techniques to detect potential adulterants in coconut water through classification models. The dataset comprises infrared spectra from coconut water samples spiked with 15 different types of potential sugar substitutes, including: sugars, artificial sweeteners, and sugar alcohols. The interaction of infrared light with molecular bonds generates unique molecular fingerprints, forming the basis of our analysis. Departing from previous research predominantly reliant on linear-based chemometrics for adulterant detection, our study explored linear, non-linear, and combined feature extraction models. By developing an interactive application utilizing principal component analysis (PCA) and t-distributed stochastic neighbor embedding (t-SNE), non-targeted sugar adulterant detection was streamlined through enhanced visualization and pattern recognition. Targeted analysis using ensemble learning random forest (RF) and deep learning 1-dimensional convolutional neural network (1D CNN) achieved higher classification accuracies (95% and 96%, respectively) compared to sparse partial least squares discriminant analysis (sPLS-DA) at 77% and support vector machine (SVM) at 88% on the same dataset. The CNN\'s demonstrated classification accuracy is complemented by exceptional efficiency through its ability to train and test on raw data.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    背景:低肌肉质量和骨骼肌质量(SMM)损失与患者不良预后相关,但是手动SMM量化的耗时性质禁止在临床实践中实施该指标.因此,与人工定量相比,我们评估了自动SMM定量的可行性.我们评估了低肌肉质量的诊断准确性以及SMM(变化)与结直肠癌(CRC)患者生存率的关联。
    方法:分析两项临床研究中纳入的CRC患者的计算机断层扫描(CT)图像。我们比较了i)手动与自动分割第三腰椎[L3]椎骨的预选切片(“半自动”),和ii)手动L3-切片选择+手动分割与自动L3切片选择+自动分割(“全自动”)。使用Quantib身体成分v0.2.1进行自动L3选择和自动分割。Bland-Altman分析,受试者内变异系数(WSCV)和组内相关系数(ICC)用于评估手动和自动分割之间的一致性。通过手动评估作为“金标准”来计算低肌肉质量的诊断准确性(由已确定的肌肉减少症临界值定义)。使用手动或自动评估,Cox比例风险比(HRs)用于研究一线转移性CRC治疗期间SMM变化(>5%下降是/否)与根据预后因素调整的死亡率之间的关联。SMM变化也在体重稳定的情况下单独评估(<5%,即隐匿性SMM损失)患者。
    结果:总计,分析了1580例CT扫描,而在全自动比较中分析了307次扫描的子集.纳入患者(n=553)的平均年龄为63±9岁,39%为女性。半自动比较显示偏差为-2.41cm2,95%的一致性极限[-9.02至4.20],2.25%的WSCV,ICC为0.99(95%置信区间(CI)0.97至1.00)。全自动比较方法显示偏差为-0.08cm2[-10.91至10.75],WSCV为2.85%,ICC为0.98(95%CI为0.98至0.99)。半自动比较对低肌肉质量的敏感性和特异性分别为0.99和0.89,全自动比较为0.96和0.90。在手动和自动评估中,SMM降低与较短的生存期相关(n=78/280,HR1.36[95%CI1.03至1.80]和n=89/280,HR1.38[95%CI1.05至1.81])。在人工评估中,隐匿性SMM丢失与较短的生存期相关,但在自动评估中不显著(n=44/263,HR1.43[95%CI1.01至2.03]和n=51/2639,HR1.23[95%CI0.87至1.74])。
    结论:基于深度学习的L3SMM评估显示出可靠的性能,能够使用CT措施来指导临床决策。在临床实践中实施有助于识别可能从生活方式干预中受益的低肌肉质量或(隐匿性)SMM损失的患者。
    BACKGROUND: Low muscle mass and skeletal muscle mass (SMM) loss are associated with adverse patient outcomes, but the time-consuming nature of manual SMM quantification prohibits implementation of this metric in clinical practice. Therefore, we assessed the feasibility of automated SMM quantification compared to manual quantification. We evaluated both diagnostic accuracy for low muscle mass and associations of SMM (change) with survival in colorectal cancer (CRC) patients.
    METHODS: Computed tomography (CT) images from CRC patients enrolled in two clinical studies were analyzed. We compared i) manual vs. automated segmentation of preselected slices at the third lumbar [L3] vertebra (\"semi-automated\"), and ii) manual L3-slice-selection + manual segmentation vs. automated L3-slice-selection + automated segmentation (\"fully-automated\"). Automated L3-selection and automated segmentation was performed with Quantib Body Composition v0.2.1. Bland-Altman analyses, within-subject coefficients of variation (WSCVs) and Intraclass Correlation Coefficients (ICCs) were used to evaluate the agreement between manual and automatic segmentation. Diagnostic accuracy for low muscle mass (defined by an established sarcopenia cut-off) was calculated with manual assessment as the \"gold standard\". Using either manual or automated assessment, Cox proportional hazard ratios (HRs) were used to study the association between changes in SMM (>5% decrease yes/no) during first-line metastatic CRC treatment and mortality adjusted for prognostic factors. SMM change was also assessed separately in weight-stable (<5%, i.e. occult SMM loss) patients.
    RESULTS: In total, 1580 CT scans were analyzed, while a subset of 307 scans were analyzed in the fully-automated comparison. Included patients (n = 553) had a mean age of 63 ± 9 years and 39% were female. The semi-automated comparison revealed a bias of -2.41 cm2, 95% limits of agreement [-9.02 to 4.20], a WSCV of 2.25%, and an ICC of 0.99 (95% confidence intervals (CI) 0.97 to 1.00). The fully-automated comparison method revealed a bias of -0.08 cm2 [-10.91 to 10.75], a WSCV of 2.85% and an ICC of 0.98 (95% CI 0.98 to 0.99). Sensitivity and specificity for low muscle mass were 0.99 and 0.89 for the semi-automated comparison and 0.96 and 0.90 for the fully-automated comparison. SMM decrease was associated with shorter survival in both manual and automated assessment (n = 78/280, HR 1.36 [95% CI 1.03 to 1.80] and n = 89/280, HR 1.38 [95% CI 1.05 to 1.81]). Occult SMM loss was associated with shorter survival in manual assessment, but not significantly in automated assessment (n = 44/263, HR 1.43 [95% CI 1.01 to 2.03] and n = 51/2639, HR 1.23 [95% CI 0.87 to 1.74]).
    CONCLUSIONS: Deep-learning based assessment of SMM at L3 shows reliable performance, enabling the use of CT measures to guide clinical decision making. Implementation in clinical practice helps to identify patients with low muscle mass or (occult) SMM loss who may benefit from lifestyle interventions.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    计算线虫是一项劳动密集型且耗时的任务,然而,它是各种定量线虫研究的关键步骤;准备盆栽中的初始种群密度和最终种群密度,与管理相关的不同目标的微绘图和田间试验,包括线虫侵染病灶的采样和位置。线虫学家长期以来一直在与线虫计数的复杂性作斗争,导致了几项旨在自动化这一过程的研究举措。然而,这些研究工作主要集中在识别单个图像中的单类对象。为了增强这项技术的实用性,迫切需要一种算法,该算法不仅可以同时检测多类对象,而且可以对其进行分类。本研究致力于通过开发包含多种深度学习算法的用户友好的图形用户界面(GUI)来应对这一挑战。允许同时识别和分类线虫卵和Meloidogynespp的第二阶段少年。使用两种不同的成像系统总共生成了650张鸡蛋图像和1339张青少年图像,产生8655个鸡蛋和4742个使用边界框和分割进行注释的Meloidogyne少年,分别。深度学习模型是通过利用称为YOLOv8x的卷积神经网络(CNN)机器学习架构开发的。我们的结果表明,在94%和93%的实例中,这些模型正确地将卵识别为卵,将Meloidogyne幼体识别为Meloidogyne幼体,分别。该模型在模型预测与未见图像的观察之间显示出高于0.70的系数相关性。我们的研究展示了这些模型在未来实际应用中的潜在实用性。GUI通过作者的GitHub存储库(https://github.com/bresilla/nematode_counting)免费提供给公众。虽然这项研究目前集中在一个属,有计划扩大GUI的能力,包括其他具有经济意义的植物寄生线虫属。实现这些目标,包括提高不同成像系统的模型精度,可能需要多个线虫团队和实验室之间的合作,而不是单一实体的工作。随着线虫学家对利用机器学习的兴趣日益浓厚,作者对所有人都可以使用的通用自动线虫计数系统的潜在开发充满信心。本文旨在作为启动实现这一重要目标的全球合作的框架和催化剂。
    Counting nematodes is a labor-intensive and time-consuming task, yet it is a pivotal step in various quantitative nematological studies; preparation of initial population densities and final population densities in pot, micro-plot and field trials for different objectives related to management including sampling and location of nematode infestation foci. Nematologists have long battled with the complexities of nematode counting, leading to several research initiatives aimed at automating this process. However, these research endeavors have primarily focused on identifying single-class objects within individual images. To enhance the practicality of this technology, there\'s a pressing need for an algorithm that cannot only detect but also classify multiple classes of objects concurrently. This study endeavors to tackle this challenge by developing a user-friendly Graphical User Interface (GUI) that comprises multiple deep learning algorithms, allowing simultaneous recognition and categorization of nematode eggs and second stage juveniles of Meloidogyne spp. In total of 650 images for eggs and 1339 images for juveniles were generated using two distinct imaging systems, resulting in 8655 eggs and 4742 Meloidogyne juveniles annotated using bounding box and segmentation, respectively. The deep-learning models were developed by leveraging the Convolutional Neural Networks (CNNs) machine learning architecture known as YOLOv8x. Our results showed that the models correctly identified eggs as eggs and Meloidogyne juveniles as Meloidogyne juveniles in 94% and 93% of instances, respectively. The model demonstrated higher than 0.70 coefficient correlation between model predictions and observations on unseen images. Our study has showcased the potential utility of these models in practical applications for the future. The GUI is made freely available to the public through the author\'s GitHub repository (https://github.com/bresilla/nematode_counting). While this study currently focuses on one genus, there are plans to expand the GUI\'s capabilities to include other economically significant genera of plant parasitic nematodes. Achieving these objectives, including enhancing the models\' accuracy on different imaging systems, may necessitate collaboration among multiple nematology teams and laboratories, rather than being the work of a single entity. With the increasing interest among nematologists in harnessing machine learning, the authors are confident in the potential development of a universal automated nematode counting system accessible to all. This paper aims to serve as a framework and catalyst for initiating global collaboration toward this important goal.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    准确识别手绘化学结构对于在传统实验室笔记本中数字化手写化学信息或促进在平板电脑或智能手机上输入基于手写笔的结构至关重要。然而,手绘结构的固有可变性对现有的光化学结构识别(OCSR)软件提出了挑战。为了解决这个问题,我们提出了一种增强的化学图识别(DECIMER)架构,利用卷积神经网络(CNN)和变压器的组合来提高手绘化学结构的识别。该模型包含一个EfficientNetV2CNN编码器,该编码器从手绘图像中提取特征,接下来是一个转换器解码器,它将提取的特征转换为简化分子输入线输入系统(SMILES)字符串。我们的模型是使用RanDepict生成的合成手绘图像进行训练的,用于描绘具有不同样式元素的化学结构的工具。使用手绘化学结构的真实世界数据集进行基准测试以评估模型的性能。结果表明,与其他方法相比,我们改进的DECIMER体系结构显示出显着提高的识别准确性。科学贡献:这里介绍的新DECIMER模型完善了我们以前的研究工作,是目前唯一专门为识别手绘化学结构而定制的开源模型。增强的模型在处理手写样式的变化方面表现更好,线条粗细,和背景噪音,使其适用于现实世界的应用。DECIMER手绘结构识别模型及其源代码已在许可下作为开放源代码包提供。
    Accurate recognition of hand-drawn chemical structures is crucial for digitising hand-written chemical information in traditional laboratory notebooks or facilitating stylus-based structure entry on tablets or smartphones. However, the inherent variability in hand-drawn structures poses challenges for existing Optical Chemical Structure Recognition (OCSR) software. To address this, we present an enhanced Deep lEarning for Chemical ImagE Recognition (DECIMER) architecture that leverages a combination of Convolutional Neural Networks (CNNs) and Transformers to improve the recognition of hand-drawn chemical structures. The model incorporates an EfficientNetV2 CNN encoder that extracts features from hand-drawn images, followed by a Transformer decoder that converts the extracted features into Simplified Molecular Input Line Entry System (SMILES) strings. Our models were trained using synthetic hand-drawn images generated by RanDepict, a tool for depicting chemical structures with different style elements. A benchmark was performed using a real-world dataset of hand-drawn chemical structures to evaluate the model\'s performance. The results indicate that our improved DECIMER architecture exhibits a significantly enhanced recognition accuracy compared to other approaches. SCIENTIFIC CONTRIBUTION: The new DECIMER model presented here refines our previous research efforts and is currently the only open-source model tailored specifically for the recognition of hand-drawn chemical structures. The enhanced model performs better in handling variations in handwriting styles, line thicknesses, and background noise, making it suitable for real-world applications. The DECIMER hand-drawn structure recognition model and its source code have been made available as an open-source package under a permissive license.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    目的:比较深度学习加速全身(WB)与常规扩散序列的图像质量。
    方法:连续50例骨髓癌患者接受WB-MRI检查。两位专家比较了轴向b900s/mm2和相应的深分辨率增强(DRB)加速扩散加权成像(DWI)序列(采集时间:6:42分钟)与常规序列(采集时间:14分钟)的最大强度投影(MIP)。读者评估配对图像的噪声,人工制品,信号脂肪抑制,和使用李克特量表的病变显著性,也表达了他们的整体主观偏好。统计比较正常组织和癌病灶的信噪比和对比噪声比(SNR和CNR)以及表观扩散系数(ADC)值。
    结果:总体而言,在几乎80%的患者中,放射科医生首选轴向DRBb900和/或相应的MIP图像,特别是高体重指数(BMI>25kg/m2)的患者。在定性评估中,在56-100%的病例中,轴向DRB图像是首选(首选/强烈首选),而DRBMIP图像在52-96%的病例中受到青睐。所有正常组织中DRB-SNR/CNR均较高(p<0.05)。对于癌症病变,DRB-SNR较高(p<0.001),但CNR并没有什么不同。大脑和腰大肌的DRB-ADC值明显更高,但不适用于癌症病变(平均差:+53µm2/s)。类间相关系数分析显示良好至优异的一致性(95%CI0.75-0.93)。
    结论:DRB序列产生更高质量的轴向DWI,从而改善MIP并显著减少采集时间。然而,需要考虑正常组织ADC值的差异。
    结论:深度学习加速扩散序列以减少的采集时间产生高质量的轴向图像和MIP。这种进步可以使全身MRI用于评估骨髓癌患者的更多采用。
    结论:深度学习重建能够使WB扩散序列的采集时间减少50%以上。在近80%的病例中,由于伪影较少,放射科医生更喜欢DRB图像。改进的背景信号抑制,更高的信噪比,体重指数较高的患者的病变显著性增加。来自DRB图像的癌症病变扩散率与常规序列没有不同。
    OBJECTIVE: To compare the image quality of deep learning accelerated whole-body (WB) with conventional diffusion sequences.
    METHODS: Fifty consecutive patients with bone marrow cancer underwent WB-MRI. Two experts compared axial b900 s/mm2 and the corresponding maximum intensity projections (MIP) of deep resolve boost (DRB) accelerated diffusion-weighted imaging (DWI) sequences (time of acquisition: 6:42 min) against conventional sequences (time of acquisition: 14 min). Readers assessed paired images for noise, artefacts, signal fat suppression, and lesion conspicuity using Likert scales, also expressing their overall subjective preference. Signal-to-noise and contrast-to-noise ratios (SNR and CNR) and the apparent diffusion coefficient (ADC) values of normal tissues and cancer lesions were statistically compared.
    RESULTS: Overall, radiologists preferred either axial DRB b900 and/or corresponding MIP images in almost 80% of the patients, particularly in patients with a high body-mass index (BMI > 25 kg/m2). In qualitative assessments, axial DRB images were preferred (preferred/strongly preferred) in 56-100% of cases, whereas DRB MIP images were favoured in 52-96% of cases. DRB-SNR/CNR was higher in all normal tissues (p < 0.05). For cancer lesions, the DRB-SNR was higher (p < 0.001), but the CNR was not different. DRB-ADC values were significantly higher for the brain and psoas muscles, but not for cancer lesions (mean difference: + 53 µm2/s). Inter-class correlation coefficient analysis showed good to excellent agreement (95% CI 0.75-0.93).
    CONCLUSIONS: DRB sequences produce higher-quality axial DWI, resulting in improved MIPs and significantly reduced acquisition times. However, differences in the ADC values of normal tissues need to be considered.
    CONCLUSIONS: Deep learning accelerated diffusion sequences produce high-quality axial images and MIP at reduced acquisition times. This advancement could enable the increased adoption of Whole Body-MRI for the evaluation of patients with bone marrow cancer.
    CONCLUSIONS: Deep learning reconstruction enables a more than 50% reduction in acquisition time for WB diffusion sequences. DRB images were preferred by radiologists in almost 80% of cases due to fewer artefacts, improved background signal suppression, higher signal-to-noise ratio, and increased lesion conspicuity in patients with higher body mass index. Cancer lesion diffusivity from DRB images was not different from conventional sequences.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    背景:对于剂量学,对全身SPECT/CT成像的需求,双头愤怒相机需要较长的采集时间,正在增加。在这里,我们评估了稀疏获取的投影,并评估了添加深度学习生成的合成中间投影(SIP)是否可以在保持剂量测定准确性的同时提高图像质量。
    方法:本研究包括16例患者,用177Lu-DOTATATE进行SPECT/CT成像(120个投影,120P)在四个时间点。设计并训练深度神经网络(CUSIP),以从30个获得的投影(30P)中编译90个SIP。120P,30P,并使用基于蒙特卡洛的OSEM重建重建了三个不同的CUSIP集(30P90SIP)(产生120P_rec,30P_rec,和CUSIP_recs)。视觉比较噪声水平。归一化均方根误差的定量测量,归一化平均绝对误差,峰值信噪比,和结构相似性进行了评估,对每个重建组的肾脏和骨髓吸收剂量进行估算。
    结果:使用SIP在视觉上改善了噪声水平。所有定量测量都显示出CUSIP集和120P之间的高度相似性。线性回归显示,所有重建装置的肾脏和骨髓吸收剂量几乎完全一致,与120P_rec的剂量相比(R2≥0.97)。与120P_rec相比,肾脏吸收剂量的平均相对差异,对于所有重建集,在3%以内。对于骨髓吸收剂量,相对差异有更高的耗散,CUSIP_recs的平均相对差异优于30P_rec(4%以内,9%)。30P_rec的肾脏和骨髓吸收剂量与120_rec的有统计学意义。与最佳表现的CUSIP_rec的吸收剂量相反,没有发现统计学上的显著差异。
    结论:进行SPECT/CT重建时,使用SIP可以大大减少SPECT/CT成像中的采集持续时间,能够以令人满意的剂量精度采集高图像质量的多个视场。
    BACKGROUND: For dosimetry, the demand for whole-body SPECT/CT imaging, which require long acquisition durations with dual-head Anger cameras, is increasing. Here we evaluated sparsely acquired projections and assessed whether the addition of deep-learning-generated synthetic intermediate projections (SIPs) could improve the image quality while preserving dosimetric accuracy.
    METHODS: This study included 16 patients treated with 177Lu-DOTATATE with SPECT/CT imaging (120 projections, 120P) at four time points. Deep neural networks (CUSIPs) were designed and trained to compile 90 SIPs from 30 acquired projections (30P). The 120P, 30P, and three different CUSIP sets (30P + 90 SIPs) were reconstructed using Monte Carlo-based OSEM reconstruction (yielding 120P_rec, 30P_rec, and CUSIP_recs). The noise levels were visually compared. Quantitative measures of normalised root mean square error, normalised mean absolute error, peak signal-to-noise ratio, and structural similarity were evaluated, and kidney and bone marrow absorbed doses were estimated for each reconstruction set.
    RESULTS: The use of SIPs visually improved noise levels. All quantitative measures demonstrated high similarity between CUSIP sets and 120P. Linear regression showed nearly perfect concordance of the kidney and bone marrow absorbed doses for all reconstruction sets, compared to the doses of 120P_rec (R2 ≥ 0.97). Compared to 120P_rec, the mean relative difference in kidney absorbed dose, for all reconstruction sets, was within 3%. For bone marrow absorbed doses, there was a higher dissipation in relative differences, and CUSIP_recs outperformed 30P_rec in mean relative difference (within 4% compared to 9%). Kidney and bone marrow absorbed doses for 30P_rec were statistically significantly different from those of 120_rec, as opposed to the absorbed doses of the best performing CUSIP_rec, where no statistically significant difference was found.
    CONCLUSIONS: When performing SPECT/CT reconstruction, the use of SIPs can substantially reduce acquisition durations in SPECT/CT imaging, enabling acquisition of multiple fields of view of high image quality with satisfactory dosimetric accuracy.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

公众号