BERT

BERT
  • 文章类型: Journal Article
    早期的癌症检测和治疗取决于发现导致癌症的特定基因。遗传突变的分类最初是手动完成的。然而,这个过程依赖于病理学家,可能是一项耗时的任务。因此,为了提高临床解释的精度,研究人员开发了利用下一代测序技术进行自动化突变分析的计算算法.本文利用四个深度学习分类模型和生物医学文本的训练集合。这些模型包括来自生物医学文本挖掘变压器(BioBERT)的双向编码器表示,为生物上下文实现的专用语言模型。在多个任务中令人印象深刻的结果,包括文本分类,语言推理,和问题回答,可以通过简单地添加一个额外的层到Biobert模型获得。此外,来自变压器(BERT)的双向编码器表示,长短期记忆(LSTM),和双向LSTM(BiLSTM)已被利用在基于文本证据对基因突变进行分类方面产生非常好的结果。工作中使用的数据集是由纪念斯隆·凯特琳癌症中心(MSKCC)创建的,其中包含几个突变。此外,该数据集在Kaggle研究预测竞赛中构成了重大分类挑战。在开展工作中,确定了三个挑战:巨大的文本长度,数据的偏见表示,和重复的数据实例。根据常用的评估指标,实验结果表明,BioBERT模型优于其他模型,F1得分为0.87和0.850MCC,与使用BERT模型获得的F1评分为0.70的文献中的类似结果相比,这可以被认为是改进的性能。
    Early cancer detection and treatment depend on the discovery of specific genes that cause cancer. The classification of genetic mutations was initially done manually. However, this process relies on pathologists and can be a time-consuming task. Therefore, to improve the precision of clinical interpretation, researchers have developed computational algorithms that leverage next-generation sequencing technologies for automated mutation analysis. This paper utilized four deep learning classification models with training collections of biomedical texts. These models comprise bidirectional encoder representations from transformers for Biomedical text mining (BioBERT), a specialized language model implemented for biological contexts. Impressive results in multiple tasks, including text classification, language inference, and question answering, can be obtained by simply adding an extra layer to the BioBERT model. Moreover, bidirectional encoder representations from transformers (BERT), long short-term memory (LSTM), and bidirectional LSTM (BiLSTM) have been leveraged to produce very good results in categorizing genetic mutations based on textual evidence. The dataset used in the work was created by Memorial Sloan Kettering Cancer Center (MSKCC), which contains several mutations. Furthermore, this dataset poses a major classification challenge in the Kaggle research prediction competitions. In carrying out the work, three challenges were identified: enormous text length, biased representation of the data, and repeated data instances. Based on the commonly used evaluation metrics, the experimental results show that the BioBERT model outperforms other models with an F1 score of 0.87 and 0.850 MCC, which can be considered as improved performance compared to similar results in the literature that have an F1 score of 0.70 achieved with the BERT model.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    情感分析近年来已成为自然语言处理研究的关键领域。该研究旨在比较各种情绪分析技术的性能,包括基于词典的,机器学习,Bi-LSTM,BERT,和GPT-3方法,使用两个常用的数据集,IMDB评论和感受140。目的是确定样本数据集的最佳性能技术,与2021年世界卫生组织烟草控制框架公约缔约方第九次会议(COP9)相关的推文。
    进行两阶段评价。在第一阶段,使用标准评估指标,如准确性,F1分数,和精度。在第二阶段,从第一阶段开始的最佳性能技术被应用于部分注释的COP9会议相关推文.
    在第一阶段,BERT获得了最高的F1分数(IMDB为0.9380,Sentiment140为0.8114),其次是GPT-3(0.9119和0.7913)和Bi-LSTM(0.8971和0.7778)。在第二阶段,GPT-3在部分注释的COP9会议相关推文上表现最好,F1得分为0.8812。
    该研究证明了BERT和GPT-3等预训练模型对情感分析任务的有效性,在标准数据集上优于传统技术。此外,GPT-3在部分注释的COP9推文上的更好性能突出了它能够很好地推广到具有有限注释的特定领域数据。这为研究人员和从业人员提供了一个可行的选择,即在不同领域的数据有限或没有注释数据的情况下,使用预训练模型进行情感分析。
    UNASSIGNED: Sentiment analysis has become a crucial area of research in natural language processing in recent years. The study aims to compare the performance of various sentiment analysis techniques, including lexicon-based, machine learning, Bi-LSTM, BERT, and GPT-3 approaches, using two commonly used datasets, IMDB reviews and Sentiment140. The objective is to identify the best-performing technique for an exemplar dataset, tweets associated with the WHO Framework Convention on Tobacco Control Ninth Conference of the Parties in 2021 (COP9).
    UNASSIGNED: A two-stage evaluation was conducted. In the first stage, various techniques were compared on standard sentiment analysis datasets using standard evaluation metrics such as accuracy, F1-score, and precision. In the second stage, the best-performing techniques from the first stage were applied to partially annotated COP9 conference-related tweets.
    UNASSIGNED: In the first stage, BERT achieved the highest F1-scores (0.9380 for IMDB and 0.8114 for Sentiment 140), followed by GPT-3 (0.9119 and 0.7913) and Bi-LSTM (0.8971 and 0.7778). In the second stage, GPT-3 performed the best for sentiment analysis on partially annotated COP9 conference-related tweets, with an F1-score of 0.8812.
    UNASSIGNED: The study demonstrates the effectiveness of pre-trained models like BERT and GPT-3 for sentiment analysis tasks, outperforming traditional techniques on standard datasets. Moreover, the better performance of GPT-3 on the partially annotated COP9 tweets highlights its ability to generalize well to domain-specific data with limited annotations. This provides researchers and practitioners with a viable option of using pre-trained models for sentiment analysis in scenarios with limited or no annotated data across different domains.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    背景:在本文中,我们提出了一种自动的文章分类方法,利用大型语言模型(LLM)的强大功能。
    目的:本研究的目的是根据科学眼科论文的文本内容评估各种LLM的适用性。
    方法:我们开发了一种基于自然语言处理技术的模型,包括高级LLM,对科技论文的文本内容进行处理和分析。具体来说,我们使用零镜头学习LLM,并将双向和自回归变压器(BART)及其变体与来自变压器的双向编码器表示(BERT)及其变体进行了比较,比如distilbert,Scibert,PubmedBERT,Biobert要评估LLM,我们汇编了1000篇与眼部疾病相关的文章的数据集(视网膜疾病[RenD]),由6名专家组成的小组熟练地将其注释为19个不同的类别。除了文章的分类,我们还对不同的分类组进行了分析,以发现该领域的模式和趋势。
    结果:分类结果证明了LLM在没有人为干预的情况下对大量眼科论文进行分类的有效性。基于RenD数据集,该模型实现了0.86的平均准确度和0.85的平均F1得分。
    结论:所提出的框架在准确性和效率上都取得了显著的提高。它在眼科领域的应用展示了其知识组织和检索的潜力。我们进行了趋势分析,使研究人员和临床医生能够轻松地对相关论文进行分类和检索,在文献综述和信息收集以及不同学科中新兴科学趋势的识别方面节省时间和精力。此外,该模型在其他科学领域的可扩展性扩大了其在促进跨不同学科的研究和趋势分析方面的影响。
    BACKGROUND: In this paper, we present an automated method for article classification, leveraging the power of large language models (LLMs).
    OBJECTIVE: The aim of this study is to evaluate the applicability of various LLMs based on textual content of scientific ophthalmology papers.
    METHODS: We developed a model based on natural language processing techniques, including advanced LLMs, to process and analyze the textual content of scientific papers. Specifically, we used zero-shot learning LLMs and compared Bidirectional and Auto-Regressive Transformers (BART) and its variants with Bidirectional Encoder Representations from Transformers (BERT) and its variants, such as distilBERT, SciBERT, PubmedBERT, and BioBERT. To evaluate the LLMs, we compiled a data set (retinal diseases [RenD] ) of 1000 ocular disease-related articles, which were expertly annotated by a panel of 6 specialists into 19 distinct categories. In addition to the classification of articles, we also performed analysis on different classified groups to find the patterns and trends in the field.
    RESULTS: The classification results demonstrate the effectiveness of LLMs in categorizing a large number of ophthalmology papers without human intervention. The model achieved a mean accuracy of 0.86 and a mean F1-score of 0.85 based on the RenD data set.
    CONCLUSIONS: The proposed framework achieves notable improvements in both accuracy and efficiency. Its application in the domain of ophthalmology showcases its potential for knowledge organization and retrieval. We performed a trend analysis that enables researchers and clinicians to easily categorize and retrieve relevant papers, saving time and effort in literature review and information gathering as well as identification of emerging scientific trends within different disciplines. Moreover, the extendibility of the model to other scientific fields broadens its impact in facilitating research and trend analysis across diverse disciplines.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    背景:研究空白是指现有知识体系中未回答的问题,由于缺乏研究或结果不确定。研究差距是科学研究的重要起点和动力。确定研究差距的传统方法,如文献综述和专家意见,可能很耗时,劳动密集型,而且容易产生偏见.在处理快速发展或时间敏感的主题时,它们也可能不足。因此,需要创新的可扩展方法来确定研究差距,系统地评估文献,并优先考虑感兴趣的主题的进一步研究领域。
    目的:在本文中,我们提出了一种基于机器学习的方法,通过分析科学文献来识别研究差距。我们使用COVID-19大流行作为案例研究。
    方法:我们使用COVID-19开放研究(CORD-19)数据集进行了分析,以确定COVID-19文献中的研究空白,其中包括1,121,433篇与COVID-19大流行有关的论文。我们的方法基于BERTopic主题建模技术,它利用转换器和基于类的术语频率-逆文档频率来创建密集的集群,从而允许易于解释的主题。我们基于BERTopic的方法涉及3个阶段:嵌入文档,聚类文档(降维和聚类),和代表主题(生成候选和最大化候选相关性)。
    结果:应用研究选择标准后,我们在本研究的分析中纳入了33,206篇摘要.最终的研究差距清单确定了21个不同的领域,分为6个主要主题。这些主题是:\“COVID-19的病毒”,\“COVID-19的危险因素”,\“预防COVID-19”,\“COVID-19的治疗”,\“COVID-19期间的医疗保健服务,\”和COVID-19的影响。\"最突出的话题,在超过一半的分析研究中观察到,是“COVID-19的影响。
    结论:提出的基于机器学习的方法有可能发现科学文献中的研究空白。本研究并非旨在取代选定主题内的个别文献研究。相反,它可以作为指导,在与以前的出版物指定用于未来探索的研究问题相关的特定领域制定精确的文献检索查询。未来的研究应该利用从目标区域最常见的数据库中检索到的最新研究列表。在可行的情况下,全文或,至少,应该对讨论部分进行分析,而不是将其分析局限于摘要。此外,未来的研究可以评估更有效的建模算法,尤其是那些将主题建模与统计不确定性量化相结合的方法,如共形预测。
    BACKGROUND: Research gaps refer to unanswered questions in the existing body of knowledge, either due to a lack of studies or inconclusive results. Research gaps are essential starting points and motivation in scientific research. Traditional methods for identifying research gaps, such as literature reviews and expert opinions, can be time consuming, labor intensive, and prone to bias. They may also fall short when dealing with rapidly evolving or time-sensitive subjects. Thus, innovative scalable approaches are needed to identify research gaps, systematically assess the literature, and prioritize areas for further study in the topic of interest.
    OBJECTIVE: In this paper, we propose a machine learning-based approach for identifying research gaps through the analysis of scientific literature. We used the COVID-19 pandemic as a case study.
    METHODS: We conducted an analysis to identify research gaps in COVID-19 literature using the COVID-19 Open Research (CORD-19) data set, which comprises 1,121,433 papers related to the COVID-19 pandemic. Our approach is based on the BERTopic topic modeling technique, which leverages transformers and class-based term frequency-inverse document frequency to create dense clusters allowing for easily interpretable topics. Our BERTopic-based approach involves 3 stages: embedding documents, clustering documents (dimension reduction and clustering), and representing topics (generating candidates and maximizing candidate relevance).
    RESULTS: After applying the study selection criteria, we included 33,206 abstracts in the analysis of this study. The final list of research gaps identified 21 different areas, which were grouped into 6 principal topics. These topics were: \"virus of COVID-19,\" \"risk factors of COVID-19,\" \"prevention of COVID-19,\" \"treatment of COVID-19,\" \"health care delivery during COVID-19,\" \"and impact of COVID-19.\" The most prominent topic, observed in over half of the analyzed studies, was \"the impact of COVID-19.\"
    CONCLUSIONS: The proposed machine learning-based approach has the potential to identify research gaps in scientific literature. This study is not intended to replace individual literature research within a selected topic. Instead, it can serve as a guide to formulate precise literature search queries in specific areas associated with research questions that previous publications have earmarked for future exploration. Future research should leverage an up-to-date list of studies that are retrieved from the most common databases in the target area. When feasible, full texts or, at minimum, discussion sections should be analyzed rather than limiting their analysis to abstracts. Furthermore, future studies could evaluate more efficient modeling algorithms, especially those combining topic modeling with statistical uncertainty quantification, such as conformal prediction.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Multicenter Study
    背景:非结构化格式的电子健康记录(EHR)是临床和生物医学领域研究的宝贵信息来源。然而,在这些记录可用于研究目的之前,在某些情况下,必须删除敏感的健康信息(SHI),以保护患者的隐私。基于规则和基于机器学习的方法已被证明在去识别方面是有效的。然而,很少有研究研究了基于转换器的语言模型和规则的组合。
    目的:本研究的目的是使用规则和转换器为澳大利亚EHR文本注释开发混合去识别管道。该研究还旨在研究预训练单词嵌入和基于转换器的语言模型的影响。
    方法:在本研究中,我们提出了一种称为OpenDeID的混合去识别管道,它是使用澳大利亚基于多中心EHR的语料库OpenDeID语料库开发的。OpenDeID语料库由2100个病理学报告组成,其中有来自1833名患者的38,414个SHI实体。OpenDeID管道结合了关联规则的混合方法,有监督的深度学习,和预先训练的语言模型。
    结果:通过微调DischargeSummaryBiobert模型并结合各种预处理和后处理规则,OpenDeID获得了0.9659的最佳F1得分。OpenDeID管道已部署在大型三级教学医院,并实时处理了8000多个非结构化EHR文本注释。
    结论:OpenDeID管道是一种混合去标识管道,用于去标识非结构化EHR文本注释中的SHI实体。该管道已在大型多中心语料库上进行了评估。外部验证将作为我们未来工作的一部分进行,以评估OpenDeID管道的有效性。
    Electronic health records (EHRs) in unstructured formats are valuable sources of information for research in both the clinical and biomedical domains. However, before such records can be used for research purposes, sensitive health information (SHI) must be removed in several cases to protect patient privacy. Rule-based and machine learning-based methods have been shown to be effective in deidentification. However, very few studies investigated the combination of transformer-based language models and rules.
    The objective of this study is to develop a hybrid deidentification pipeline for Australian EHR text notes using rules and transformers. The study also aims to investigate the impact of pretrained word embedding and transformer-based language models.
    In this study, we present a hybrid deidentification pipeline called OpenDeID, which is developed using an Australian multicenter EHR-based corpus called OpenDeID Corpus. The OpenDeID corpus consists of 2100 pathology reports with 38,414 SHI entities from 1833 patients. The OpenDeID pipeline incorporates a hybrid approach of associative rules, supervised deep learning, and pretrained language models.
    The OpenDeID achieved a best F1-score of 0.9659 by fine-tuning the Discharge Summary BioBERT model and incorporating various preprocessing and postprocessing rules. The OpenDeID pipeline has been deployed at a large tertiary teaching hospital and has processed over 8000 unstructured EHR text notes in real time.
    The OpenDeID pipeline is a hybrid deidentification pipeline to deidentify SHI entities in unstructured EHR text notes. The pipeline has been evaluated on a large multicenter corpus. External validation will be undertaken as part of our future work to evaluate the effectiveness of the OpenDeID pipeline.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    基于Transformer的解决自然语言处理(NLP)任务(如BERT和GPT)的方法由于其实现高性能的能力而越来越受欢迎。这些方法受益于使用巨大的数据大小来创建预先训练的模型以及理解句子中单词上下文的能力。它们在信息检索领域的使用被认为可以提高有效性和效率。本文演示了一种基于BERT的方法(CASBRT)实现,用于在使用本体综合注释的数据上构建搜索工具。数据是使用Physiome模型库(PMR)中的CellML标准编写的生物模拟模型的集合。生物模拟模型在结构上由常量和变量的基本实体组成,这些实体构造了更高级别的实体,例如组件,reactions,和模型。找到这些特定于其级别的实体对于有关变量重用的各种目的是有益的,实验设置,和模型审计。最初,我们为常量和变量搜索(最低级别实体)创建了表示复合注释实体的嵌入。然后,这些低级实体嵌入被垂直和有效地组合在一起,以创建更高级的实体嵌入来搜索组件,模型,images,和模拟设置。我们的方法是一般的,因此,它可以用于创建搜索工具,其中包含其他语义注释的数据-以SBML格式编码的生物模拟模型,例如。我们的工具被命名为生物模拟模型搜索引擎(BMSE)。
    The Transformer-based approaches to solving natural language processing (NLP) tasks such as BERT and GPT are gaining popularity due to their ability to achieve high performance. These approaches benefit from using enormous data sizes to create pre-trained models and the ability to understand the context of words in a sentence. Their use in the information retrieval domain is thought to increase effectiveness and efficiency. This paper demonstrates a BERT-based method (CASBERT) implementation to build a search tool over data annotated compositely using ontologies. The data was a collection of biosimulation models written using the CellML standard in the Physiome Model Repository (PMR). A biosimulation model structurally consists of basic entities of constants and variables that construct higher-level entities such as components, reactions, and the model. Finding these entities specific to their level is beneficial for various purposes regarding variable reuse, experiment setup, and model audit. Initially, we created embeddings representing compositely-annotated entities for constant and variable search (lowest level entity). Then, these low-level entity embeddings were vertically and efficiently combined to create higher-level entity embeddings to search components, models, images, and simulation setups. Our approach was general, so it can be used to create search tools with other data semantically annotated with ontologies - biosimulation models encoded in the SBML format, for example. Our tool is named Biosimulation Model Search Engine (BMSE).
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    循证医学,医疗保健专业人员在做出决策时参考最佳可用证据的做法,构成现代医疗保健的基础。然而,它依赖于劳动密集型的系统审查,领域专家必须从成千上万的出版物中汇总和提取信息,主要是随机对照试验(RCT)结果,进入证据表。本文通过将问题分解为两个语言处理任务来研究自动化证据表生成:命名实体识别,它识别文本中的关键实体,如药物名称,和关系提取,映射它们的关系,将它们分成有序的元组。我们专注于从已发表的RCT摘要中自动制表句子,这些摘要报告了研究结果。作为联合提取管道的一部分,开发了两个深度神经网络模型,使用迁移学习和基于转换器的语言表示的原理。为了训练和测试这些模型,开发了一个新的黄金标准语料库,包括来自六个疾病领域的550多个结果句。这种方法显示出显著的优势,我们的系统在多个自然语言处理任务和疾病领域表现良好,以及推广到训练期间看不见的疾病领域。此外,我们表明,这些结果是可以通过在170例句上训练我们的模型来实现的。最终的系统是一个概念证明,证据表的生成可以是半自动的,代表着完全自动化系统审查的一步。
    Evidence-based medicine, the practice in which healthcare professionals refer to the best available evidence when making decisions, forms the foundation of modern healthcare. However, it relies on labour-intensive systematic reviews, where domain specialists must aggregate and extract information from thousands of publications, primarily of randomised controlled trial (RCT) results, into evidence tables. This paper investigates automating evidence table generation by decomposing the problem across two language processing tasks: named entity recognition, which identifies key entities within text, such as drug names, and relation extraction, which maps their relationships for separating them into ordered tuples. We focus on the automatic tabulation of sentences from published RCT abstracts that report the results of the study outcomes. Two deep neural net models were developed as part of a joint extraction pipeline, using the principles of transfer learning and transformer-based language representations. To train and test these models, a new gold-standard corpus was developed, comprising over 550 result sentences from six disease areas. This approach demonstrated significant advantages, with our system performing well across multiple natural language processing tasks and disease areas, as well as in generalising to disease domains unseen during training. Furthermore, we show these results were achievable through training our models on as few as 170 example sentences. The final system is a proof of concept that the generation of evidence tables can be semi-automated, representing a step towards fully automating systematic reviews.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    背景:机构名称的规范化对于文献检索非常重要,学术成就统计,和评价研究机构的竞争力。作者写作习惯和拼写错误的差异导致机构的不同名称,这影响了出版数据的分析。随着深度学习模型的发展和自然语言处理方法的日益成熟,训练基于深度学习的机构名称规范化模型可以提高机构名称规范化在语义层面的准确性。
    目的:本研究旨在基于多源文献中隶属关系数据的特征融合,为机构名称规范化训练基于深度学习的模型,借助权限文件实现机构名称变体的规范化,经过几轮训练和优化,达到较高的规范准确率。
    方法:在本研究中,基于来自变压器(BERT)的双向编码器表示和其他深度学习模型训练了面向机构名称规范化的模型,包括机构分类模型,机构层次关系提取模型,以及机构匹配和合并模型。然后对模型进行训练,通过预训练和微调自动学习机构特征,从3个数据库的隶属关系数据中提取机构名称,完成标准化过程:维度,WebofScience,还有Scopus.
    结果:发现训练后的模型至少可以实现3种功能。首先,该模型可以识别与权限文件一致的机构名称,并通过唯一的机构ID将该名称与文件相关联。第二,它可以识别非标准机构名称变体,如单数形式,复数变化,和缩写,并更新权限文件。第三,它可以识别未注册的机构并将其添加到授权文件中,所以当这个机构再次出现时,该模型可以识别并将其视为注册机构。此外,检验结果表明,归一化模型的准确率达到93.79%,表明该模型在机构名称规范化方面表现良好。
    结论:本研究中训练的基于深度学习的机构名称规范化模型具有很高的准确性。因此,它可以广泛应用于研究机构竞争力的评估,对机构研究领域的分析,建立机构间合作网络,其中,具有较高的应用价值。
    BACKGROUND: The normalization of institution names is of great importance for literature retrieval, statistics of academic achievements, and evaluation of the competitiveness of research institutions. Differences in authors\' writing habits and spelling mistakes lead to various names of institutions, which affects the analysis of publication data. With the development of deep learning models and the increasing maturity of natural language processing methods, training a deep learning-based institution name normalization model can increase the accuracy of institution name normalization at the semantic level.
    OBJECTIVE: This study aimed to train a deep learning-based model for institution name normalization based on the feature fusion of affiliation data from multisource literature, which would realize the normalization of institution name variants with the help of authority files and achieve a high specification accuracy after several rounds of training and optimization.
    METHODS: In this study, an institution name normalization-oriented model was trained based on bidirectional encoder representations from transformers (BERT) and other deep learning models, including the institution classification model, institutional hierarchical relation extraction model, and institution matching and merging model. The model was then trained to automatically learn institutional features by pretraining and fine-tuning, and institution names were extracted from the affiliation data of 3 databases to complete the normalization process: Dimensions, Web of Science, and Scopus.
    RESULTS: It was found that the trained model could achieve at least 3 functions. First, the model could identify the institution name that is consistent with the authority files and associate the name with the files through the unique institution ID. Second, it could identify the nonstandard institution name variants, such as singular forms, plural changes, and abbreviations, and update the authority files. Third, it could identify the unregistered institutions and add them to the authority files, so that when the institution appeared again, the model could identify and regard it as a registered institution. Moreover, the test results showed that the accuracy of the normalization model reached 93.79%, indicating the promising performance of the model for the normalization of institution names.
    CONCLUSIONS: The deep learning-based institution name normalization model trained in this study exhibited high accuracy. Therefore, it could be widely applied in the evaluation of the competitiveness of research institutions, analysis of research fields of institutions, and construction of interinstitutional cooperation networks, among others, showing high application value.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Letter
    暂无摘要。
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    在中国,内卷,这意味着压力要超越其他小组成员,在微博上引起了公众的关注。内卷的新在线内涵增强了青年群体凝聚力。与其他危机不同,这种批评也与群体凝聚力概念密切相关。然而,以前很少有与群体凝聚力相关的研究关注这一关键概念。这项研究解释了年轻人在面对内卷时为什么以及如何在网上创造团体凝聚力。首先,通过考察内卷和群体凝聚力之间的关系。第二,通过调查代沟来检查青年是否团结在网络讨论中。在此之后,这项研究分析了不同的观点,以确定为什么会出现这种群体凝聚力,年轻人如何看待内卷,以及为什么他们把“老年人”视为其他人。最后,这项研究分析了年轻人如何使用标签来吸引更多的年轻人发表意见,从而导致更大的团体凝聚力。
    通过将前沿计算方法与因果关系和轴向编码相结合,这项研究提出了一种新的方法来深入分析社交媒体上的群体凝聚力。
    结果表明,对合引发在线群体凝聚力差,与在线对卷相关的热点问题引发了基于身份的群体凝聚力。此外,年轻人比老年人更消极,他们的表达充满了基于身份的建构。通过强调社会根源并指责“其他”(老年群体),青年在线团结在一起。这些发现表明,确实存在代沟,年轻人通过“揭示社会身份”和“定位和成为”策略发布相关标签,在社交媒体上团结起来。
    研究结果强调,内卷与不良的群体凝聚力有关,社交媒体提供了一种新的方式来应对内卷危机。年轻人将使用标签来团结和指责想象中的敌人,比如老年人和上层阶级。这些发现可能有助于理解导致更多群体凝聚力的干预措施。
    UNASSIGNED: In China, involution, which means pressure to out-compete other group members, has attracted public attention on Weibo. The new online connotation of involution empowered group cohesion among youth. Dissimilar to other crises, this crise also closely relates to group cohesion concept. However, few previous group cohesion-related studies focus on this critical concept. This study explains why and how youth created group cohesion online when facing involution. First, by examining the relationship between involution and group cohesion. Second, by examining whether youth are united in the online discussion of involution by investigating the generational gap. Following this, this study analyzes the different opinions to identify why this group cohesion occurs, how youth think about involution, and why they regard \"older adults\" as others. Lastly, this study analyzes how youth use hashtags to attract more youth to voice their opinions, consequently leading to greater group cohesion.
    UNASSIGNED: By combining frontier computational methods with causation and axial coding, this study proposes a new way to in-depth analyze group cohesion on social media.
    UNASSIGNED: The results indicate that involution triggers poor online group cohesion, and online involution-related hot issues trigger identity-based group cohesion. Additionally, youth are significantly more negative than older adults, and their expressions are full of identity-based construction. By stressing the social roots and blaming the \"other\" (older adult group), youth united together online. These findings indicated that a generation gap does indeed exist and that youth unite on social media by posting related hashtags via \"revealing social identity\" and \"positioning and becoming\" strategies.
    UNASSIGNED: The findings stress that involution is related to poor group cohesion and that social media offers a new way to face the involution crisis. Youth will use hashtags to unite and blame imagined enemies, such as older adults and the upper class. These findings might assist in understanding interventions that lead to more group cohesion.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

公众号