BERT

BERT
  • 文章类型: Journal Article
    转移性乳腺癌(MBC)仍然是女性癌症相关死亡的主要原因。这项工作介绍了一种创新的非侵入性乳腺癌分类模型,旨在改善癌症转移的识别。虽然这项研究标志着预测MBC的初步探索,额外的调查对于验证MBC的发生至关重要.我们的方法结合了大型语言模型(LLM)的优势,特别是来自变压器(BERT)模型的双向编码器表示,图神经网络(GNN)的强大功能,可根据组织病理学报告预测MBC患者。本文介绍了一种用于转移性乳腺癌预测(BG-MBC)的BERT-GNN方法,该方法集成了从BERT模型得出的图形信息。在这个模型中,节点是根据病人的医疗记录构建的,虽然BERT嵌入被用来对组织病理学报告中的单词进行矢量化表示,从而通过采用三种不同的方法(即单变量选择,用于特征重要性的额外树分类器,和Shapley值,以确定影响最显著的特征)。确定在模型训练期间作为嵌入生成的676个中最关键的30个特征,我们的模型进一步增强了其预测能力。BG-MBC模型具有出色的准确性,在识别MBC患者时,检出率为0.98,曲线下面积(AUC)为0.98。这种显著的表现归功于模型对LLM从组织病理学报告中产生的注意力得分的利用,有效地捕获相关特征进行分类。
    Metastatic breast cancer (MBC) continues to be a leading cause of cancer-related deaths among women. This work introduces an innovative non-invasive breast cancer classification model designed to improve the identification of cancer metastases. While this study marks the initial exploration into predicting MBC, additional investigations are essential to validate the occurrence of MBC. Our approach combines the strengths of large language models (LLMs), specifically the bidirectional encoder representations from transformers (BERT) model, with the powerful capabilities of graph neural networks (GNNs) to predict MBC patients based on their histopathology reports. This paper introduces a BERT-GNN approach for metastatic breast cancer prediction (BG-MBC) that integrates graph information derived from the BERT model. In this model, nodes are constructed from patient medical records, while BERT embeddings are employed to vectorise representations of the words in histopathology reports, thereby capturing semantic information crucial for classification by employing three distinct approaches (namely univariate selection, extra trees classifier for feature importance, and Shapley values to identify the features that have the most significant impact). Identifying the most crucial 30 features out of 676 generated as embeddings during model training, our model further enhances its predictive capabilities. The BG-MBC model achieves outstanding accuracy, with a detection rate of 0.98 and an area under curve (AUC) of 0.98, in identifying MBC patients. This remarkable performance is credited to the model\'s utilisation of attention scores generated by the LLM from histopathology reports, effectively capturing pertinent features for classification.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    毒性鉴定在维护人类健康中起着关键作用,因为它可以提醒人类长期接触各种化合物所造成的潜在危害。确定毒性的实验方法很耗时,而且昂贵,而计算方法为早期识别毒性提供了一种替代方法。例如,一些经典的ML和DL方法,在毒性预测中表现出优异的性能。然而,这些方法也有一些缺陷,例如过度依赖人工特征和容易过度拟合,等。提出具有优越预测性能的新模型仍然是一项紧迫的任务。在这项研究中,我们提出了一种基于motifs级图的多视图预训练语言模型,叫做3MTox,用于毒性鉴定。3MTox模型使用来自变压器的双向编码器表示(BERT)作为骨干框架,和一个图案图作为输入。大量实验的结果表明,我们的3MTox模型在毒性基准数据集上实现了最先进的性能,并且优于所考虑的基准模型。此外,模型的可解释性保证了它能快速准确地识别给定分子中的毒性位点,从而有助于确定毒性状态和相关分析。我们认为3MTox模型是目前可用于毒性鉴定的最有前途的工具之一。
    Toxicity identification plays a key role in maintaining human health, as it can alert humans to the potential hazards caused by long-term exposure to a wide variety of chemical compounds. Experimental methods for determining toxicity are time-consuming, and costly, while computational methods offer an alternative for the early identification of toxicity. For example, some classical ML and DL methods, which demonstrate excellent performance in toxicity prediction. However, these methods also have some defects, such as over-reliance on artificial features and easy overfitting, etc. Proposing novel models with superior prediction performance is still an urgent task. In this study, we propose a motifs-level graph-based multi-view pretraining language model, called 3MTox, for toxicity identification. The 3MTox model uses Bidirectional Encoder Representations from Transformers (BERT) as the backbone framework, and a motif graph as input. The results of extensive experiments showed that our 3MTox model achieved state-of-the-art performance on toxicity benchmark datasets and outperformed the baseline models considered. In addition, the interpretability of the model ensures that the it can quickly and accurately identify toxicity sites in a given molecule, thereby contributing to the determination of the status of toxicity and associated analyses. We think that the 3MTox model is among the most promising tools that are currently available for toxicity identification.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    在全球化的浪潮中,文化融合现象激增,突出强调跨文化交际中固有的挑战。为了应对这些挑战,当代研究已将重点转移到人机对话上。尤其是在人机对话的教育范式中,分析用户对话中的情感识别尤为重要。准确识别和理解用户的情感倾向以及人机交互和游戏的效率和体验。本研究旨在提高人机对话中的语言情感识别能力。它提出了一种基于来自变压器(BERT)的双向编码器表示的混合模型(BCBA),卷积神经网络(CNN),双向门控递归单位(BiGRU),注意机制。该模型利用BERT模型从文本中提取语义和句法特征。同时,它集成了CNN和BiGRU网络,以更深入地研究文本特征,增强模型在细致入微的情感识别方面的熟练程度。此外,通过引入注意力机制,该模型可以根据单词的情绪倾向为单词分配不同的权重。这使其能够优先考虑具有可辨别的情绪倾向的单词,以进行更精确的情绪分析。通过在两个数据集上的实验验证,BCBA模型在情感识别和分类任务中取得了显著的效果。该模型的准确性和F1得分都有了显著提高,平均准确率为0.84,平均F1评分为0.8。混淆矩阵分析揭示了该模型的最小分类错误率。此外,随着迭代次数的增加,模型的召回率稳定在约0.7。这一成就展示了该模型在语义理解和情感分析方面的强大功能,并展示了其在跨文化背景下处理语言表达中的情感特征方面的优势。本研究提出的BCBA模型为人机对话中的情感识别提供了有效的技术支持,这对于构建更加智能、人性化的人机交互系统具有重要意义。在未来,我们将继续优化模型的结构,提高其处理复杂情绪和跨语言情绪识别的能力,并探索将该模型应用于更多的实际场景,进一步促进人机对话技术的发展和应用。
    Amid the wave of globalization, the phenomenon of cultural amalgamation has surged in frequency, bringing to the fore the heightened prominence of challenges inherent in cross-cultural communication. To address these challenges, contemporary research has shifted its focus to human-computer dialogue. Especially in the educational paradigm of human-computer dialogue, analysing emotion recognition in user dialogues is particularly important. Accurately identify and understand users\' emotional tendencies and the efficiency and experience of human-computer interaction and play. This study aims to improve the capability of language emotion recognition in human-computer dialogue. It proposes a hybrid model (BCBA) based on bidirectional encoder representations from transformers (BERT), convolutional neural networks (CNN), bidirectional gated recurrent units (BiGRU), and the attention mechanism. This model leverages the BERT model to extract semantic and syntactic features from the text. Simultaneously, it integrates CNN and BiGRU networks to delve deeper into textual features, enhancing the model\'s proficiency in nuanced sentiment recognition. Furthermore, by introducing the attention mechanism, the model can assign different weights to words based on their emotional tendencies. This enables it to prioritize words with discernible emotional inclinations for more precise sentiment analysis. The BCBA model has achieved remarkable results in emotion recognition and classification tasks through experimental validation on two datasets. The model has significantly improved both accuracy and F1 scores, with an average accuracy of 0.84 and an average F1 score of 0.8. The confusion matrix analysis reveals a minimal classification error rate for this model. Additionally, as the number of iterations increases, the model\'s recall rate stabilizes at approximately 0.7. This accomplishment demonstrates the model\'s robust capabilities in semantic understanding and sentiment analysis and showcases its advantages in handling emotional characteristics in language expressions within a cross-cultural context. The BCBA model proposed in this study provides effective technical support for emotion recognition in human-computer dialogue, which is of great significance for building more intelligent and user-friendly human-computer interaction systems. In the future, we will continue to optimize the model\'s structure, improve its capability in handling complex emotions and cross-lingual emotion recognition, and explore applying the model to more practical scenarios to further promote the development and application of human-computer dialogue technology.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    癌症免疫学为传统癌症治疗提供了新的选择,如放疗和化疗。一个值得注意的替代方案是开发基于癌症新抗原的个性化疫苗。此外,变形金刚被认为是人工智能的革命性发展,对自然语言处理(NLP)任务产生重大影响,近年来已被用于蛋白质组学研究。在这种情况下,我们进行了系统的文献综述,以研究变形金刚如何应用于新抗原检测过程的每个阶段.此外,我们绘制了当前的管道,并检查了涉及癌症疫苗的临床试验结果。
    Cancer immunology offers a new alternative to traditional cancer treatments, such as radiotherapy and chemotherapy. One notable alternative is the development of personalized vaccines based on cancer neoantigens. Moreover, Transformers are considered a revolutionary development in artificial intelligence with a significant impact on natural language processing (NLP) tasks and have been utilized in proteomics studies in recent years. In this context, we conducted a systematic literature review to investigate how Transformers are applied in each stage of the neoantigen detection process. Additionally, we mapped current pipelines and examined the results of clinical trials involving cancer vaccines.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    目标:随着医疗保健环境中有效的PPM分诊的需求和工作量的增加,患者门户消息(PPM)激增,刺激了对AI驱动解决方案的探索,以简化医疗保健工作流程。确保及时响应患者以满足他们的医疗保健需求。然而,在PPMs中,人们较少关注隔离和理解患者的主要关注点,这种做法有可能产生更细致入微的见解,并提高医疗保健服务和以患者为中心的护理质量.
    方法:我们提出了一种融合框架,通过卷积神经网络利用具有不同语言优势的预训练语言模型(LM),通过多类分类精确识别患者的主要关注点。我们研究了3种传统的机器学习模型,9个基于BERT的语言模型,6个融合模型,和2合奏模型。
    结果:我们的实验结果强调了基于BERT的模型与传统机器学习模型相比所取得的卓越性能。值得注意的是,我们的融合模型成为性能最好的解决方案,显著提高了总体平均准确率77.67±2.74%和F1评分74.37±3.70%.
    结论:本研究强调了用于患者主要关注检测的多类别分类的可行性和有效性,以及用于增强主要关注检测的拟议融合框架。
    结论:通过融合多个预先训练的LM来增强多类分类的使用不仅提高了PPM中患者主要关注点识别的准确性和效率,而且有助于管理不断增长的PPM在医疗保健中的数量。确保及时和准确地解决关键的病人沟通。
    OBJECTIVE: The surge in patient portal messages (PPMs) with increasing needs and workloads for efficient PPM triage in healthcare settings has spurred the exploration of AI-driven solutions to streamline the healthcare workflow processes, ensuring timely responses to patients to satisfy their healthcare needs. However, there has been less focus on isolating and understanding patient primary concerns in PPMs-a practice which holds the potential to yield more nuanced insights and enhances the quality of healthcare delivery and patient-centered care.
    METHODS: We propose a fusion framework to leverage pretrained language models (LMs) with different language advantages via a Convolution Neural Network for precise identification of patient primary concerns via multi-class classification. We examined 3 traditional machine learning models, 9 BERT-based language models, 6 fusion models, and 2 ensemble models.
    RESULTS: The outcomes of our experimentation underscore the superior performance achieved by BERT-based models in comparison to traditional machine learning models. Remarkably, our fusion model emerges as the top-performing solution, delivering a notably improved accuracy score of 77.67 ± 2.74% and an F1 score of 74.37 ± 3.70% in macro-average.
    CONCLUSIONS: This study highlights the feasibility and effectiveness of multi-class classification for patient primary concern detection and the proposed fusion framework for enhancing primary concern detection.
    CONCLUSIONS: The use of multi-class classification enhanced by a fusion of multiple pretrained LMs not only improves the accuracy and efficiency of patient primary concern identification in PPMs but also aids in managing the rising volume of PPMs in healthcare, ensuring critical patient communications are addressed promptly and accurately.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    推理解决是自然语言处理中的关键任务。很难评估大跨度文本的相似性,这使得文本级编码有些挑战。本文首先比较了常用的方法来提高模型的全局信息收集能力对BERT编码性能的影响。基于此,为了提高BERT编码模型在不同文本跨度下的适用性,设计了多尺度上下文信息模块。此外,通过尺寸扩展提高线性可分性。最后,使用交叉熵损失作为损失函数。在本文设计的模块中添加BERT和spanBERT后,F1分别增加了0.5%和0.2%,分别。
    Coreference resolution is a key task in Natural Language Processing. It is difficult to evaluate the similarity of long-span texts, which makes text-level encoding somewhat challenging. This paper first compares the impact of commonly used methods to improve the global information collection ability of the model on the BERT encoding performance. Based on this, a multi-scale context information module is designed to improve the applicability of the BERT encoding model under different text spans. In addition, improving linear separability through dimension expansion. Finally, cross-entropy loss is used as the loss function. After adding BERT and span BERT to the module designed in this article, F1 increased by 0.5% and 0.2%, respectively.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    随着自然语言处理(NLP)的快速发展,预训练语言模型(PLM),如BERT、Biobert,ChatGPT在各种医学NLP任务中显示出巨大的潜力。本文调查了将PLM应用于各种医学NLP任务的前沿成就。具体来说,我们首先简要介绍PLMS,概述PLMS在医学中的研究。接下来,我们对医学NLP中的任务类型进行分类和讨论,涵盖文本摘要,问答,机器翻译,情绪分析,命名实体识别,信息提取,医学教育,关系提取,和文本挖掘。对于每种类型的任务,我们首先提供基本概念的概述,主要方法,应用PLM的优势,应用PLM应用程序的基本步骤,用于培训和测试的数据集,以及任务评估的指标。随后,总结了最近的重要研究成果,分析他们的动机,优势与劣势,相似性与差异性,讨论潜在的限制。此外,我们通过比较被审查论文的引文数和发表论文的会议和期刊的声誉和影响来评估本文所审查研究的质量和影响力。通过这些指标,我们进一步确定了当前最关注的研究课题。最后,我们期待着未来的研究方向,包括增强模型的可靠性,可解释性,和公平,促进PLMs在临床实践中的应用。此外,本次调查还收集了一些模型代码和相关数据集的下载链接,这对于在医学中应用NLP技术的研究人员和寻求通过AI技术增强其专业知识和医疗保健服务的医疗专业人员来说是有价值的参考。
    With the rapid progress in Natural Language Processing (NLP), Pre-trained Language Models (PLM) such as BERT, BioBERT, and ChatGPT have shown great potential in various medical NLP tasks. This paper surveys the cutting-edge achievements in applying PLMs to various medical NLP tasks. Specifically, we first brief PLMS and outline the research of PLMs in medicine. Next, we categorise and discuss the types of tasks in medical NLP, covering text summarisation, question-answering, machine translation, sentiment analysis, named entity recognition, information extraction, medical education, relation extraction, and text mining. For each type of task, we first provide an overview of the basic concepts, the main methodologies, the advantages of applying PLMs, the basic steps of applying PLMs application, the datasets for training and testing, and the metrics for task evaluation. Subsequently, a summary of recent important research findings is presented, analysing their motivations, strengths vs weaknesses, similarities vs differences, and discussing potential limitations. Also, we assess the quality and influence of the research reviewed in this paper by comparing the citation count of the papers reviewed and the reputation and impact of the conferences and journals where they are published. Through these indicators, we further identify the most concerned research topics currently. Finally, we look forward to future research directions, including enhancing models\' reliability, explainability, and fairness, to promote the application of PLMs in clinical practice. In addition, this survey also collect some download links of some model codes and the relevant datasets, which are valuable references for researchers applying NLP techniques in medicine and medical professionals seeking to enhance their expertise and healthcare service through AI technology.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    自然语言处理(NLP)中的许多最新结果依赖于大型预训练语言模型(PLM)。这些模型由大量参数组成,这些参数使用大量训练数据进行调整。这些因素导致模型记忆部分训练数据,使他们容易受到各种隐私攻击。这令人担忧,特别是当这些模型应用于临床领域时,数据非常敏感。训练数据假名化是旨在缓解这些问题的隐私保护技术。此技术会自动识别敏感实体,并将其替换为现实但不敏感的代理。在先前的研究中,假名化已产生了有希望的结果。然而,以前没有研究对PLM的训练前数据和用于解决临床NLP任务的微调数据应用假名.这项研究评估了针对五个临床NLP任务进行微调的瑞典临床BERT模型的端到端假名化预测性能的影响。进行了大量的统计检验,在使用假名微调数据时显示对性能的最小危害。结果也没有发现预训练和微调数据的端到端假名化的恶化。这些结果表明,可以在不损害训练PLM的数据效用的情况下,对训练数据进行假名化以降低隐私风险。
    Many state-of-the-art results in natural language processing (NLP) rely on large pre-trained language models (PLMs). These models consist of large amounts of parameters that are tuned using vast amounts of training data. These factors cause the models to memorize parts of their training data, making them vulnerable to various privacy attacks. This is cause for concern, especially when these models are applied in the clinical domain, where data are very sensitive. Training data pseudonymization is a privacy-preserving technique that aims to mitigate these problems. This technique automatically identifies and replaces sensitive entities with realistic but non-sensitive surrogates. Pseudonymization has yielded promising results in previous studies. However, no previous study has applied pseudonymization to both the pre-training data of PLMs and the fine-tuning data used to solve clinical NLP tasks. This study evaluates the effects on the predictive performance of end-to-end pseudonymization of Swedish clinical BERT models fine-tuned for five clinical NLP tasks. A large number of statistical tests are performed, revealing minimal harm to performance when using pseudonymized fine-tuning data. The results also find no deterioration from end-to-end pseudonymization of pre-training and fine-tuning data. These results demonstrate that pseudonymizing training data to reduce privacy risks can be done without harming data utility for training PLMs.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    早期的癌症检测和治疗取决于发现导致癌症的特定基因。遗传突变的分类最初是手动完成的。然而,这个过程依赖于病理学家,可能是一项耗时的任务。因此,为了提高临床解释的精度,研究人员开发了利用下一代测序技术进行自动化突变分析的计算算法.本文利用四个深度学习分类模型和生物医学文本的训练集合。这些模型包括来自生物医学文本挖掘变压器(BioBERT)的双向编码器表示,为生物上下文实现的专用语言模型。在多个任务中令人印象深刻的结果,包括文本分类,语言推理,和问题回答,可以通过简单地添加一个额外的层到Biobert模型获得。此外,来自变压器(BERT)的双向编码器表示,长短期记忆(LSTM),和双向LSTM(BiLSTM)已被利用在基于文本证据对基因突变进行分类方面产生非常好的结果。工作中使用的数据集是由纪念斯隆·凯特琳癌症中心(MSKCC)创建的,其中包含几个突变。此外,该数据集在Kaggle研究预测竞赛中构成了重大分类挑战。在开展工作中,确定了三个挑战:巨大的文本长度,数据的偏见表示,和重复的数据实例。根据常用的评估指标,实验结果表明,BioBERT模型优于其他模型,F1得分为0.87和0.850MCC,与使用BERT模型获得的F1评分为0.70的文献中的类似结果相比,这可以被认为是改进的性能。
    Early cancer detection and treatment depend on the discovery of specific genes that cause cancer. The classification of genetic mutations was initially done manually. However, this process relies on pathologists and can be a time-consuming task. Therefore, to improve the precision of clinical interpretation, researchers have developed computational algorithms that leverage next-generation sequencing technologies for automated mutation analysis. This paper utilized four deep learning classification models with training collections of biomedical texts. These models comprise bidirectional encoder representations from transformers for Biomedical text mining (BioBERT), a specialized language model implemented for biological contexts. Impressive results in multiple tasks, including text classification, language inference, and question answering, can be obtained by simply adding an extra layer to the BioBERT model. Moreover, bidirectional encoder representations from transformers (BERT), long short-term memory (LSTM), and bidirectional LSTM (BiLSTM) have been leveraged to produce very good results in categorizing genetic mutations based on textual evidence. The dataset used in the work was created by Memorial Sloan Kettering Cancer Center (MSKCC), which contains several mutations. Furthermore, this dataset poses a major classification challenge in the Kaggle research prediction competitions. In carrying out the work, three challenges were identified: enormous text length, biased representation of the data, and repeated data instances. Based on the commonly used evaluation metrics, the experimental results show that the BioBERT model outperforms other models with an F1 score of 0.87 and 0.850 MCC, which can be considered as improved performance compared to similar results in the literature that have an F1 score of 0.70 achieved with the BERT model.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    大型语言模型(LLM)分析和响应自由书写文本的能力在精神病学领域引起了越来越多的兴奋;此类模型的应用为精神病学应用带来了独特的机遇和挑战。这篇综述文章旨在全面概述精神病学中的LLM,他们的模型架构,潜在的用例,和临床考虑。诸如ChatGPT/GPT-4之类的LLM框架是针对大量文本数据进行训练的,这些文本数据有时会针对特定任务进行微调。这开辟了广泛的可能的精神病学应用,例如准确预测特定疾病的个体患者风险因素,从事治疗干预,分析治疗材料,仅举几例。然而,在精神病学环境中收养会带来许多挑战,包括LLM的固有限制和偏见,对可解释性和隐私的担忧,以及产生的错误信息造成的潜在损害。这篇综述涵盖了潜在的机会和局限性,并强调了在现实世界的精神病学背景下应用这些模型时的潜在考虑因素。
    The ability of Large Language Models (LLMs) to analyze and respond to freely written text is causing increasing excitement in the field of psychiatry; the application of such models presents unique opportunities and challenges for psychiatric applications. This review article seeks to offer a comprehensive overview of LLMs in psychiatry, their model architecture, potential use cases, and clinical considerations. LLM frameworks such as ChatGPT/GPT-4 are trained on huge amounts of text data that are sometimes fine-tuned for specific tasks. This opens up a wide range of possible psychiatric applications, such as accurately predicting individual patient risk factors for specific disorders, engaging in therapeutic intervention, and analyzing therapeutic material, to name a few. However, adoption in the psychiatric setting presents many challenges, including inherent limitations and biases in LLMs, concerns about explainability and privacy, and the potential damage resulting from produced misinformation. This review covers potential opportunities and limitations and highlights potential considerations when these models are applied in a real-world psychiatric context.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

公众号