transfer learning

迁移学习
  • 文章类型: Journal Article
    深度学习深刻影响了各个领域,特别是医学图像分析。该领域的传统迁移学习方法依赖于在特定领域的医学数据集上预训练的模型,这限制了它们的通用性和可访问性。在这项研究中,我们提出了一个叫做真实世界特征迁移学习的新框架,它利用最初在大规模通用数据集如ImageNet上训练的骨干模型。与从头开始训练的模型相比,我们评估了这种方法的有效性和鲁棒性,专注于对X射线图像中的肺炎进行分类的任务。我们的实验,其中包括将灰度图像转换为RGB格式,证明了真实世界的特征迁移学习在各种性能指标上始终优于传统的训练方法。这一进步有可能通过利用从通用预训练模型学习的丰富特征表示来加速医学成像中的深度学习应用。所提出的方法克服了特定领域预训练模型的局限性,从而加速医疗诊断和医疗保健领域的创新。从数学的角度来看,我们形式化现实世界的特征迁移学习的概念,并提供了一个严格的数学公式的问题。我们的实验结果提供了支持这种方法有效性的经验证据,为进一步的理论分析和探索奠定基础。这项工作有助于更广泛地理解跨域的特征可转移性,并对开发准确有效的医学图像分析模型具有重要意义。即使在资源受限的环境中。
    Deep learning has profoundly influenced various domains, particularly medical image analysis. Traditional transfer learning approaches in this field rely on models pretrained on domain-specific medical datasets, which limits their generalizability and accessibility. In this study, we propose a novel framework called real-world feature transfer learning, which utilizes backbone models initially trained on large-scale general-purpose datasets such as ImageNet. We evaluate the effectiveness and robustness of this approach compared to models trained from scratch, focusing on the task of classifying pneumonia in X-ray images. Our experiments, which included converting grayscale images to RGB format, demonstrate that real-world-feature transfer learning consistently outperforms conventional training approaches across various performance metrics. This advancement has the potential to accelerate deep learning applications in medical imaging by leveraging the rich feature representations learned from general-purpose pretrained models. The proposed methodology overcomes the limitations of domain-specific pretrained models, thereby enabling accelerated innovation in medical diagnostics and healthcare. From a mathematical perspective, we formalize the concept of real-world feature transfer learning and provide a rigorous mathematical formulation of the problem. Our experimental results provide empirical evidence supporting the effectiveness of this approach, laying the foundation for further theoretical analysis and exploration. This work contributes to the broader understanding of feature transferability across domains and has significant implications for the development of accurate and efficient models for medical image analysis, even in resource-constrained settings.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    这项研究的重点是开发基于运动图像(MI)的脑机接口(BMI),使用深度学习算法来控制下肢机器人外骨骼。该研究旨在通过利用深度学习的优势来克服传统BMI方法的局限性,如自动特征提取和迁移学习。评估BMI的实验方案设计为异步,允许受试者按照自己的意愿执行心理任务。
    共有5名身体健康的受试者参加了一系列实验。来自其中两个会话的大脑信号用于通过迁移学习开发通用的深度学习模型。随后,在剩余的课程中对该模型进行了微调,并进行了评估.比较了三种不同的深度学习方法:一种没有经过微调,另一个微调了模型的所有层,第三个只微调了最后三层。评估阶段涉及参与者使用第二种深度学习方法进行解码的神经活动对外骨骼设备的专有闭环控制。
    与基于每个受试者和实验阶段训练的空间特征的方法相比,对三种深度学习方法进行了评估。展示他们的卓越表现。有趣的是,没有微调的深度学习方法实现了与基于特征的方法相当的性能,这表明,在来自不同个体和以前会话的数据上训练的通用模型可以产生类似的效果。在三种深度学习方法中,进行了比较,微调所有层权重展示了最高的性能。
    这项研究代表了迈向未来免校准方法的第一步。尽管努力通过利用其他受试者的数据来减少校准时间,完全消除被证明是不可能实现的。这项研究的发现对推进无校准方法具有显著意义,承诺将培训试验的需求降至最低。此外,本研究中采用的实验评估方案旨在复制现实生活场景,在行走或停止步态等行为的决策中,给予参与者更高的自主权。
    This research focused on the development of a motor imagery (MI) based brain-machine interface (BMI) using deep learning algorithms to control a lower-limb robotic exoskeleton. The study aimed to overcome the limitations of traditional BMI approaches by leveraging the advantages of deep learning, such as automated feature extraction and transfer learning. The experimental protocol to evaluate the BMI was designed as asynchronous, allowing subjects to perform mental tasks at their own will.
    A total of five healthy able-bodied subjects were enrolled in this study to participate in a series of experimental sessions. The brain signals from two of these sessions were used to develop a generic deep learning model through transfer learning. Subsequently, this model was fine-tuned during the remaining sessions and subjected to evaluation. Three distinct deep learning approaches were compared: one that did not undergo fine-tuning, another that fine-tuned all layers of the model, and a third one that fine-tuned only the last three layers. The evaluation phase involved the exclusive closed-loop control of the exoskeleton device by the participants\' neural activity using the second deep learning approach for the decoding.
    The three deep learning approaches were assessed in comparison to an approach based on spatial features that was trained for each subject and experimental session, demonstrating their superior performance. Interestingly, the deep learning approach without fine-tuning achieved comparable performance to the features-based approach, indicating that a generic model trained on data from different individuals and previous sessions can yield similar efficacy. Among the three deep learning approaches compared, fine-tuning all layer weights demonstrated the highest performance.
    This research represents an initial stride toward future calibration-free methods. Despite the efforts to diminish calibration time by leveraging data from other subjects, complete elimination proved unattainable. The study\'s discoveries hold notable significance for advancing calibration-free approaches, offering the promise of minimizing the need for training trials. Furthermore, the experimental evaluation protocol employed in this study aimed to replicate real-life scenarios, granting participants a higher degree of autonomy in decision-making regarding actions such as walking or stopping gait.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    背景:一些研究人员质疑人工智能(AI)系统在系统开发过程中未考虑的人群中用于女性时是否能保持其性能。
    目的:评估迁移学习作为改善AI系统在乳腺癌检测中的泛化的一种方式的影响。
    方法:这项芬兰回顾性病例对照研究涉及191名被诊断患有乳腺癌的女性和191名匹配的健康对照。我们选择了使用大型美国数据集训练的最先进的AI系统进行乳腺癌检测。在两个实验设置中评价所选择的基线系统。首先,我们将我们的私人芬兰样本作为一个独立的测试集进行了检查,该测试集在系统的开发中没有被考虑(看不见的群体).第二,对基线系统进行了再训练,试图通过迁移学习来改善其在未见人群中的表现.为了分析性能,我们使用Delong检验的接受者工作特征曲线(AUC)下的面积。
    结果:考虑了两种版本的基线系统:ImageOnly和Heatmap。ImageOnly和Heatmaps版本在美国数据集中的平均AUC值为0.82±0.008和0.88±0.003,在未见过的人群中进行评估时,平均AUC值为0.56(95%CI=0.50-0.62)和0.72(95%CI=0.67-0.77),分别。经重新训练的系统的AUC值为0.61(95%CI=0.55-0.66)和0.69(95%CI=0.64-0.75),分别。基线系统和再训练系统之间没有统计学差异。
    结论:小样本的迁移学习在系统的推广方面没有显著的改善。
    BACKGROUND: Some researchers have questioned whether artificial intelligence (AI) systems maintain their performance when used for women from populations not considered during the development of the system.
    OBJECTIVE: To evaluate the impact of transfer learning as a way of improving the generalization of AI systems in the detection of breast cancer.
    METHODS: This retrospective case-control Finnish study involved 191 women diagnosed with breast cancer and 191 matched healthy controls. We selected a state-of-the-art AI system for breast cancer detection trained using a large US dataset. The selected baseline system was evaluated in two experimental settings. First, we examined our private Finnish sample as an independent test set that had not been considered in the development of the system (unseen population). Second, the baseline system was retrained to attempt to improve its performance in the unseen population by means of transfer learning. To analyze performance, we used areas under the receiver operating characteristic curve (AUCs) with DeLong\'s test.
    RESULTS: Two versions of the baseline system were considered: ImageOnly and Heatmaps. The ImageOnly and Heatmaps versions yielded mean AUC values of 0.82±0.008 and 0.88±0.003 in the US dataset and 0.56 (95% CI=0.50-0.62) and 0.72 (95% CI=0.67-0.77) when evaluated in the unseen population, respectively. The retrained systems achieved AUC values of 0.61 (95% CI=0.55-0.66) and 0.69 (95% CI=0.64-0.75), respectively. There was no statistical difference between the baseline system and the retrained system.
    CONCLUSIONS: Transfer learning with a small study sample did not yield a significant improvement in the generalization of the system.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    在临床氟18氟脱氧葡萄糖PET/CT报告的基础上,评估领域适应对语言模型在预测五点多维尔分数方面的性能的影响。
    作者回顾性检索了2008年至2018年在威斯康星大学麦迪逊分校机构临床影像学数据库中进行的4542份氟脱氧葡萄糖PET/CT淋巴瘤检查的文本报告和图像。在这些总报告中,1664年,从报告中提取了多维尔分数,并用作培训标签。来自变压器(BERT)模型和初始化BERT模型的双向编码器表示BioClinicalBERT,拉德伯特,和RoBERTa通过使用蒙面语言建模进行预训练而适应了核医学领域。然后在五点Deauville分数预测的任务中,将这些域适应的模型与非域适应的版本进行比较。将语言模型与视觉模型进行了比较,多模态视觉语言模型,和核医学医生,七倍蒙特卡罗交叉验证。报告了用于准确性的均值和SDs,用配对t检验的P值。
    域自适应提高了所有语言模型的性能(P=.01)。例如,领域适应后,BERT从61.3%±2.9(SD)五级精度提高到65.7%±2.2(P=0.01)。领域改编的RoBERTa(名为DARoBERTa)表现最好,达到77.4%±3.4五级精度;该模型的性能与多模式对应物(称为多模式DARoBERTa)相似(77.2%±3.2),优于最佳仅视觉模型(48.1%±3.5,P≤.001)。对数据子集进行任务的医生具有66%的五级准确率。
    领域适应提高了大型语言模型在预测PET/CT报告中的Deauville分数方面的性能。关键词淋巴瘤,PET,PET/CT,迁移学习,无监督学习,卷积神经网络(CNN)核医学,多维尔,自然语言处理,多模态学习,人工智能,机器学习,语言建模补充材料可用于本文。©RSNA,202另见本期Abajian的评注。
    UNASSIGNED: To evaluate the impact of domain adaptation on the performance of language models in predicting five-point Deauville scores on the basis of clinical fluorine 18 fluorodeoxyglucose PET/CT reports.
    UNASSIGNED: The authors retrospectively retrieved 4542 text reports and images for fluorodeoxyglucose PET/CT lymphoma examinations from 2008 to 2018 in the University of Wisconsin-Madison institutional clinical imaging database. Of these total reports, 1664 had Deauville scores that were extracted from the reports and served as training labels. The bidirectional encoder representations from transformers (BERT) model and initialized BERT models BioClinicalBERT, RadBERT, and RoBERTa were adapted to the nuclear medicine domain by pretraining using masked language modeling. These domain-adapted models were then compared with the non-domain-adapted versions on the task of five-point Deauville score prediction. The language models were compared against vision models, multimodal vision-language models, and a nuclear medicine physician, with sevenfold Monte Carlo cross-validation. Means and SDs for accuracy are reported, with P values from paired t testing.
    UNASSIGNED: Domain adaptation improved the performance of all language models (P = .01). For example, BERT improved from 61.3% ± 2.9 (SD) five-class accuracy to 65.7% ± 2.2 (P = .01) following domain adaptation. Domain-adapted RoBERTa (named DA RoBERTa) performed best, achieving 77.4% ± 3.4 five-class accuracy; this model performed similarly to its multimodal counterpart (named Multimodal DA RoBERTa) (77.2% ± 3.2) and outperformed the best vision-only model (48.1% ± 3.5, P ≤ .001). A physician given the task on a subset of the data had a five-class accuracy of 66%.
    UNASSIGNED: Domain adaptation improved the performance of large language models in predicting Deauville scores in PET/CT reports.Keywords Lymphoma, PET, PET/CT, Transfer Learning, Unsupervised Learning, Convolutional Neural Network (CNN), Nuclear Medicine, Deauville, Natural Language Processing, Multimodal Learning, Artificial Intelligence, Machine Learning, Language Modeling Supplemental material is available for this article. © RSNA, 2023See also the commentary by Abajian in this issue.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    由于单细胞RNA测序(scRNA-seq)数据中基因表达矩阵的高维性和稀疏性,再加上浅层测序产生的显著噪声,这对细胞聚类方法提出了很大的挑战。虽然已经提出了许多计算方法,现有的大多数方法都集中在处理目标数据集本身。这种方法忽视了其他物种和scRNA-seq数据批次中存在的大量知识。鉴于此,我们的论文提出了一种新的方法,称为基于图的深度嵌入聚类(GDEC),利用跨物种和批次的迁移学习。GDEC集成了图形卷积网络,有效地克服了稀疏基因表达矩阵带来的挑战。此外,DEC在GDEC中的结合使得细胞团簇在低维空间内的划分成为可能,从而减轻噪声对聚类结果的不利影响。GDEC基于现有的scRNA-seq数据集构建模型,然后应用迁移学习技术,使用从目标数据集中的有限数量的先验知识对模型进行微调。这使GDEC能够巧妙地将scRNA-seq数据跨不同的物种和批次进行聚类。通过跨物种和跨批次聚类实验,我们对GDEC和常规包装进行了比较分析。此外,我们对子宫肌瘤的scRNA-seq数据实施了GDEC.比较从Seurat包获得的结果,GDEC揭示了一种新的细胞类型(上皮细胞),并在各种细胞类型中发现了许多新的途径,从而强调了GDEC增强的分析能力。可用性和实施:https://github.com/YuzhiSun/GDEC/tree/main。
    Due to the high dimensionality and sparsity of the gene expression matrix in single-cell RNA-sequencing (scRNA-seq) data, coupled with significant noise generated by shallow sequencing, it poses a great challenge for cell clustering methods. While numerous computational methods have been proposed, the majority of existing approaches center on processing the target dataset itself. This approach disregards the wealth of knowledge present within other species and batches of scRNA-seq data. In light of this, our paper proposes a novel method named graph-based deep embedding clustering (GDEC) that leverages transfer learning across species and batches. GDEC integrates graph convolutional networks, effectively overcoming the challenges posed by sparse gene expression matrices. Additionally, the incorporation of DEC in GDEC enables the partitioning of cell clusters within a lower-dimensional space, thereby mitigating the adverse effects of noise on clustering outcomes. GDEC constructs a model based on existing scRNA-seq datasets and then applying transfer learning techniques to fine-tune the model using a limited amount of prior knowledge gleaned from the target dataset. This empowers GDEC to adeptly cluster scRNA-seq data cross different species and batches. Through cross-species and cross-batch clustering experiments, we conducted a comparative analysis between GDEC and conventional packages. Furthermore, we implemented GDEC on the scRNA-seq data of uterine fibroids. Compared results obtained from the Seurat package, GDEC unveiled a novel cell type (epithelial cells) and identified a notable number of new pathways among various cell types, thus underscoring the enhanced analytical capabilities of GDEC. Availability and implementation: https://github.com/YuzhiSun/GDEC/tree/main.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    具有完全或部分相同的内部类别的图像数据库的连续发布极大地恶化了用于真正全面的医疗诊断的自主计算机辅助诊断(CAD)系统的生产。第一个挑战是医学图像数据库的频繁大量发布,这通常有两个常见的缺点:图像复制和损坏。具有相同类别或类别的相同数据的许多后续版本没有明确的证据表明在图像数据库之间的这些相同类别的串联成功。这个问题是基于假设的实验路径上的绊脚石,用于产生可以成功地对所有这些模型进行正确分类的单一学习模型。删除冗余数据,提高性能,优化能源资源是最具挑战性的方面。在这篇文章中,我们提出了一个全球数据聚合规模模型,该模型包含从特定的全球资源中选择的六个图像数据库。建议的有效学习器基于训练任何给定数据发布中的所有独特模式,从而假设创建一个独特的数据集。HashMD5算法(MD5)为每个图像生成一个唯一的哈希值,使其适合重复删除。T分布随机邻域嵌入(t-SNE),使用可调的困惑参数,可以表示数据维度。HashMD5和t-SNE算法都是递归应用的,生成一个平衡和统一的数据库,每个类别包含相等的样本:正常,肺炎,和2019年冠状病毒病(COVID-19)。我们使用InceptionV3预训练模型和各种评估指标评估了所有建议数据和新自动化版本的性能。所提出的规模模型的性能结果显示出比传统的数据聚合更可观的结果,达到98.48%的高精度,随着高精度,召回,和F1得分。结果已通过统计t检验证明,产生t值和p值。重要的是要强调,所有的t值都是不可否认的重要,p值提供了反对零假设的无可辩驳的证据。此外,值得注意的是,当使用相同的因素诊断各种肺部感染时,Final数据集优于所有度量值的所有其他数据集。
    Continuous release of image databases with fully or partially identical inner categories dramatically deteriorates the production of autonomous Computer-Aided Diagnostics (CAD) systems for true comprehensive medical diagnostics. The first challenge is the frequent massive bulk release of medical image databases, which often suffer from two common drawbacks: image duplication and corruption. The many subsequent releases of the same data with the same classes or categories come with no clear evidence of success in the concatenation of those identical classes among image databases. This issue stands as a stumbling block in the path of hypothesis-based experiments for the production of a single learning model that can successfully classify all of them correctly. Removing redundant data, enhancing performance, and optimizing energy resources are among the most challenging aspects. In this article, we propose a global data aggregation scale model that incorporates six image databases selected from specific global resources. The proposed valid learner is based on training all the unique patterns within any given data release, thereby creating a unique dataset hypothetically. The Hash MD5 algorithm (MD5) generates a unique hash value for each image, making it suitable for duplication removal. The T-Distributed Stochastic Neighbor Embedding (t-SNE), with a tunable perplexity parameter, can represent data dimensions. Both the Hash MD5 and t-SNE algorithms are applied recursively, producing a balanced and uniform database containing equal samples per category: normal, pneumonia, and Coronavirus Disease of 2019 (COVID-19). We evaluated the performance of all proposed data and the new automated version using the Inception V3 pre-trained model with various evaluation metrics. The performance outcome of the proposed scale model showed more respectable results than traditional data aggregation, achieving a high accuracy of 98.48%, along with high precision, recall, and F1-score. The results have been proved through a statistical t-test, yielding t-values and p-values. It\'s important to emphasize that all t-values are undeniably significant, and the p-values provide irrefutable evidence against the null hypothesis. Furthermore, it\'s noteworthy that the Final dataset outperformed all other datasets across all metric values when diagnosing various lung infections with the same factors.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    迁移学习(TL)是一种尚未在医疗保健领域得到广泛研究的方法,主要应用于图像数据。本研究概述了利用个体病例安全报告(ICSR)和电子健康记录(EHR)的TL管道,用于早期检测药物不良反应(ADR),评估使用脱发和多西他赛对乳腺癌患者的使用情况。
    Transfer Learning (TL) is an approach which has not yet been widely investigated in healthcare, mostly applied in image data. This study outlines a TL pipeline leveraging Individual Case Safety reports (ICSRs) and Electronic Health Records (EHR), applied for the early detection Adverse Drug Reactions (ADR), evaluated using of alopecia and docetaxel on breast cancer patients as use case.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    工业数据稀缺是阻碍机器学习在制造业中广泛使用的最大因素之一。为了克服这个问题,迁移学习的概念得到了发展,并在最近的工业研究中受到了广泛的关注。本文重点研究了时间序列分割问题,并在末端泵测试的工业用例上首次对基于深度学习的时间序列分割进行了深入研究。特别是,我们研究是否可以通过使用来自其他领域的数据对网络进行预训练来提高深度学习模型的性能。分析了三种不同的情况:源数据和目标数据密切相关,源数据和目标数据密切相关,以及源数据和目标数据不相关。结果表明,迁移学习可以提高时间序列分割模型的准确性和训练速度。在源数据和训练数据密切相关且目标训练数据样本数量最低的情况下,可以最清楚地看到好处。然而,在非相关数据集的场景中,也观察到负迁移学习的情况。因此,这项研究强调了潜力,还有挑战,产业转移学习。
    Industrial data scarcity is one of the largest factors holding back the widespread use of machine learning in manufacturing. To overcome this problem, the concept of transfer learning was developed and has received much attention in recent industrial research. This paper focuses on the problem of time series segmentation and presents the first in-depth research on transfer learning for deep learning-based time series segmentation on the industrial use case of end-of-line pump testing. In particular, we investigate whether the performance of deep learning models can be increased by pretraining the network with data from other domains. Three different scenarios are analyzed: source and target data being closely related, source and target data being distantly related, and source and target data being non-related. The results demonstrate that transfer learning can enhance the performance of time series segmentation models with respect to accuracy and training speed. The benefit can be most clearly seen in scenarios where source and training data are closely related and the number of target training data samples is lowest. However, in the scenario of non-related datasets, cases of negative transfer learning were observed as well. Thus, the research emphasizes the potential, but also the challenges, of industrial transfer learning.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    要评估是否使用来自变压器(BERT)模型的双向编码器表示进行迁移学习,在临床语料库上预先训练,可以对自由文本放射学报告进行句子级解剖分类,即使是很少有积极例子的解剖类。
    这项回顾性研究包括2005年12月至2020年12月接受全身PET/CT成像的患者的放射学报告。这些报告中的每个句子(6272个句子)由两个注释者根据身体部位(“大脑,\"\"头颈,\"\"胸部,“\”腹部,\"\"四肢,\"\"脊柱,\"或\"其他\")。将基于BERT的迁移学习方法与两种基线机器学习方法进行了比较:双向长短期记忆(BiLSTM)和基于计数的方法。计算每种方法的精确度-召回曲线下面积(AUPRC)和接受者工作特征曲线下面积(AUC),和AUC使用DeLong检验进行比较。
    基于BERT的方法实现了0.88的宏观平均AUPRC的分类,表现优于基线。BERT的AUC结果显着高于所有类别的BiLSTM和基于计数的大脑方法的AUC结果,\"\"胸部,“\”腹部,\"和\"其他\"类(P值<.025)。BERT的AUPRC结果优于基线,即使是标签训练数据很少的类(brain:BERT,0.95,BiLSTM,0.11,基于计数,0.41;四肢:BERT,0.74,BiLSTM,0.28,基于计数,0.46;脊柱:BERT,0.82,BiLSTM,0.53,基于计数,0.69)。
    在自由文本放射学报告的句子级解剖分类中,基于BERT的迁移学习方法优于BiLSTM和基于计数的方法,即使是很少标记训练数据的解剖类。关键词:解剖学,比较研究,技术评估,迁移学习补充材料可用于本文。©RSNA,2023年。
    UNASSIGNED: To assess whether transfer learning with a bidirectional encoder representations from transformers (BERT) model, pretrained on a clinical corpus, can perform sentence-level anatomic classification of free-text radiology reports, even for anatomic classes with few positive examples.
    UNASSIGNED: This retrospective study included radiology reports of patients who underwent whole-body PET/CT imaging from December 2005 to December 2020. Each sentence in these reports (6272 sentences) was labeled by two annotators according to body part (\"brain,\" \"head & neck,\" \"chest,\" \"abdomen,\" \"limbs,\" \"spine,\" or \"others\"). The BERT-based transfer learning approach was compared with two baseline machine learning approaches: bidirectional long short-term memory (BiLSTM) and the count-based method. Area under the precision-recall curve (AUPRC) and area under the receiver operating characteristic curve (AUC) were computed for each approach, and AUCs were compared using the DeLong test.
    UNASSIGNED: The BERT-based approach achieved a macro-averaged AUPRC of 0.88 for classification, outperforming the baselines. AUC results for BERT were significantly higher than those of BiLSTM for all classes and those of the count-based method for the \"brain,\" \"chest,\" \"abdomen,\" and \"others\" classes (P values < .025). AUPRC results for BERT were superior to those of baselines even for classes with few labeled training data (brain: BERT, 0.95, BiLSTM, 0.11, count based, 0.41; limbs: BERT, 0.74, BiLSTM, 0.28, count based, 0.46; spine: BERT, 0.82, BiLSTM, 0.53, count based, 0.69).
    UNASSIGNED: The BERT-based transfer learning approach outperformed the BiLSTM and count-based approaches in sentence-level anatomic classification of free-text radiology reports, even for anatomic classes with few labeled training data.Keywords: Anatomy, Comparative Studies, Technology Assessment, Transfer Learning Supplemental material is available for this article. © RSNA, 2023.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Case Reports
    背景:提取有关传染病的相关信息是一项必不可少的任务。然而,支持公共卫生研究的一个重要障碍是缺乏有效挖掘大量健康数据的方法。
    目的:本研究旨在利用自然语言处理(NLP)提取关键信息(临床因素,健康的社会决定因素)来自文献中公布的案例。
    方法:提出的框架集成了一个数据层,用于从临床病例报告中准备数据队列;一个NLP层,用于在文本中找到临床和人口统计学命名的实体和关系;以及一个用于基准性能和分析的评估层。本研究的重点是从COVID-19病例报告中提取有价值的信息。
    结果:与基准方法相比,NLP层中的命名实体识别实现实现了约1-3%的性能增益。此外,即使没有大量的数据标签,关系提取方法在准确性方面优于基准方法(提高1-8%)。彻底的检查显示疾病的存在和患者的症状患病率。
    结论:类似的方法可以推广到其他传染病。在研究其他传染病时,使用通过迁移学习获得的先验知识是值得的。
    Extracting relevant information about infectious diseases is an essential task. However, a significant obstacle in supporting public health research is the lack of methods for effectively mining large amounts of health data.
    This study aims to use natural language processing (NLP) to extract the key information (clinical factors, social determinants of health) from published cases in the literature.
    The proposed framework integrates a data layer for preparing a data cohort from clinical case reports; an NLP layer to find the clinical and demographic-named entities and relations in the texts; and an evaluation layer for benchmarking performance and analysis. The focus of this study is to extract valuable information from COVID-19 case reports.
    The named entity recognition implementation in the NLP layer achieves a performance gain of about 1-3% compared to benchmark methods. Furthermore, even without extensive data labeling, the relation extraction method outperforms benchmark methods in terms of accuracy (by 1-8% better). A thorough examination reveals the disease\'s presence and symptoms prevalence in patients.
    A similar approach can be generalized to other infectious diseases. It is worthwhile to use prior knowledge acquired through transfer learning when researching other infectious diseases.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

公众号