Chatbots

聊天机器人
  • 文章类型: Journal Article
    随着像ChatGPT这样的大型语言模型在各个行业中的应用越来越多,它在医疗领域的潜力,特别是在标准化考试中,已成为研究的重点。
    本研究的目的是评估ChatGPT的临床表现,重点关注其在中国国家医师资格考试(CNMLE)中的准确性和可靠性。
    CNMLE2022问题集,由500个单答案多选题组成,被重新分类为15个医学亚专科。从2023年4月24日至5月15日,每个问题在OpenAI平台上用中文进行了8到12次测试。考虑了三个关键因素:GPT-3.5和4.0版本,针对医疗亚专科定制的系统角色的提示指定,为了连贯性而重复。通过准确度阈值被建立为60%。采用χ2检验和κ值评估模型的准确性和一致性。
    GPT-4.0达到了72.7%的通过精度,显著高于GPT-3.5(54%;P<.001)。GPT-4.0重复反应的变异性低于GPT-3.5(9%vs19.5%;P<.001)。然而,两个模型都显示出相对较好的响应一致性,κ值分别为0.778和0.610。系统角色在数值上提高了GPT-4.0(0.3%-3.7%)和GPT-3.5(1.3%-4.5%)的准确性,并将变异性降低了1.7%和1.8%,分别(P>0.05)。在亚组分析中,ChatGPT在不同题型之间取得了相当的准确率(P>.05)。GPT-4.0在15个亚专业中的14个超过了准确性阈值,而GPT-3.5在第一次反应的15人中有7人这样做。
    GPT-4.0通过了CNMLE,并在准确性等关键领域优于GPT-3.5,一致性,和医学专科专业知识。添加系统角色不会显着增强模型的可靠性和答案的连贯性。GPT-4.0在医学教育和临床实践中显示出有希望的潜力,值得进一步研究。
    UNASSIGNED: With the increasing application of large language models like ChatGPT in various industries, its potential in the medical domain, especially in standardized examinations, has become a focal point of research.
    UNASSIGNED: The aim of this study is to assess the clinical performance of ChatGPT, focusing on its accuracy and reliability in the Chinese National Medical Licensing Examination (CNMLE).
    UNASSIGNED: The CNMLE 2022 question set, consisting of 500 single-answer multiple choices questions, were reclassified into 15 medical subspecialties. Each question was tested 8 to 12 times in Chinese on the OpenAI platform from April 24 to May 15, 2023. Three key factors were considered: the version of GPT-3.5 and 4.0, the prompt\'s designation of system roles tailored to medical subspecialties, and repetition for coherence. A passing accuracy threshold was established as 60%. The χ2 tests and κ values were employed to evaluate the model\'s accuracy and consistency.
    UNASSIGNED: GPT-4.0 achieved a passing accuracy of 72.7%, which was significantly higher than that of GPT-3.5 (54%; P<.001). The variability rate of repeated responses from GPT-4.0 was lower than that of GPT-3.5 (9% vs 19.5%; P<.001). However, both models showed relatively good response coherence, with κ values of 0.778 and 0.610, respectively. System roles numerically increased accuracy for both GPT-4.0 (0.3%-3.7%) and GPT-3.5 (1.3%-4.5%), and reduced variability by 1.7% and 1.8%, respectively (P>.05). In subgroup analysis, ChatGPT achieved comparable accuracy among different question types (P>.05). GPT-4.0 surpassed the accuracy threshold in 14 of 15 subspecialties, while GPT-3.5 did so in 7 of 15 on the first response.
    UNASSIGNED: GPT-4.0 passed the CNMLE and outperformed GPT-3.5 in key areas such as accuracy, consistency, and medical subspecialty expertise. Adding a system role insignificantly enhanced the model\'s reliability and answer coherence. GPT-4.0 showed promising potential in medical education and clinical practice, meriting further study.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    背景:慢性乙型肝炎(CHB)在全球范围内施加了巨大的经济和社会负担。CHB的管理涉及复杂的监测和依从性挑战,特别是在中国这样的地区,CHB的高患病率与医疗保健资源限制相交。这项研究探索了ChatGPT-3.5的潜力,这是一种新兴的人工智能(AI)助手,来解决这些复杂性。在医学教育和实践方面具有显着的能力,ChatGPT-3.5的角色在管理CHB中进行了检查,特别是在医疗保健景观不同的地区。
    目的:本研究旨在揭示ChatGPT-3.5在不同语言环境中为CHB患者提供个性化医疗咨询援助方面的潜力和局限性。
    方法:问题来自已发布的指南,在线CHB社区,英文和中文搜索引擎都很完善,翻译,并汇编成96项调查。随后,这些问题在独立对话中被提交给ChatGPT-3.5和ChatGPT-4.0.然后由资深医生评估反应,注重信息化,情绪管理,重复查询的一致性,和关于医疗建议的警告声明。此外,我们采用真假问卷进一步辨别ChatGPT-3.5和ChatGPT-4.0之间封闭式问题的信息准确性差异.
    结果:来自ChatGPT-3.5的超过一半的反应(228/370,61.6%)被认为是全面的。相比之下,ChatGPT-4.0表现出更高的百分比,为74.5%(172/222;P<.001)。值得注意的是,在英语中表现优异,特别是在重复查询的信息性和一致性方面。然而,在情绪管理指导中发现了缺陷,ChatGPT-3.5中只有3.2%(6/186),ChatGPT-4.0中只有8.1%(15/154)(P=0.04)。ChatGPT-3.5在10.8%(24/222)的回复中包含免责声明,而ChatGPT-4.0在13.1%(29/222)的应答中包含免责声明(P=0.46)。当回答真假问题时,ChatGPT-4.0的准确率为93.3%(168/180),显著超过ChatGPT-3.5的准确率65.0%(117/180)(P<.001)。
    结论:在这项研究中,ChatGPT展示了作为CHB管理医疗咨询助理的基本能力。ChatGPT-3.5的工作语言的选择被认为是影响其性能的潜在因素,特别是在使用术语和口语方面,这可能会影响其在特定目标人群中的适用性。然而,作为更新的模型,ChatGPT-4.0展示了改进的信息处理能力,克服语言对信息准确性的影响。这表明,在选择大型语言模型作为医疗咨询助手时,需要考虑模型进步对应用程序的影响。鉴于这两种模型在情绪指导管理中的表现都不充分,这项研究强调了在为医疗目的部署ChatGPT时提供特定语言训练和情绪管理策略的重要性.此外,应进一步调查这些模型在对话中使用免责声明的趋势,以了解在实际应用中对患者体验的影响。
    BACKGROUND: Chronic hepatitis B (CHB) imposes substantial economic and social burdens globally. The management of CHB involves intricate monitoring and adherence challenges, particularly in regions like China, where a high prevalence of CHB intersects with health care resource limitations. This study explores the potential of ChatGPT-3.5, an emerging artificial intelligence (AI) assistant, to address these complexities. With notable capabilities in medical education and practice, ChatGPT-3.5\'s role is examined in managing CHB, particularly in regions with distinct health care landscapes.
    OBJECTIVE: This study aimed to uncover insights into ChatGPT-3.5\'s potential and limitations in delivering personalized medical consultation assistance for CHB patients across diverse linguistic contexts.
    METHODS: Questions sourced from published guidelines, online CHB communities, and search engines in English and Chinese were refined, translated, and compiled into 96 inquiries. Subsequently, these questions were presented to both ChatGPT-3.5 and ChatGPT-4.0 in independent dialogues. The responses were then evaluated by senior physicians, focusing on informativeness, emotional management, consistency across repeated inquiries, and cautionary statements regarding medical advice. Additionally, a true-or-false questionnaire was employed to further discern the variance in information accuracy for closed questions between ChatGPT-3.5 and ChatGPT-4.0.
    RESULTS: Over half of the responses (228/370, 61.6%) from ChatGPT-3.5 were considered comprehensive. In contrast, ChatGPT-4.0 exhibited a higher percentage at 74.5% (172/222; P<.001). Notably, superior performance was evident in English, particularly in terms of informativeness and consistency across repeated queries. However, deficiencies were identified in emotional management guidance, with only 3.2% (6/186) in ChatGPT-3.5 and 8.1% (15/154) in ChatGPT-4.0 (P=.04). ChatGPT-3.5 included a disclaimer in 10.8% (24/222) of responses, while ChatGPT-4.0 included a disclaimer in 13.1% (29/222) of responses (P=.46). When responding to true-or-false questions, ChatGPT-4.0 achieved an accuracy rate of 93.3% (168/180), significantly surpassing ChatGPT-3.5\'s accuracy rate of 65.0% (117/180) (P<.001).
    CONCLUSIONS: In this study, ChatGPT demonstrated basic capabilities as a medical consultation assistant for CHB management. The choice of working language for ChatGPT-3.5 was considered a potential factor influencing its performance, particularly in the use of terminology and colloquial language, and this potentially affects its applicability within specific target populations. However, as an updated model, ChatGPT-4.0 exhibits improved information processing capabilities, overcoming the language impact on information accuracy. This suggests that the implications of model advancement on applications need to be considered when selecting large language models as medical consultation assistants. Given that both models performed inadequately in emotional guidance management, this study highlights the importance of providing specific language training and emotional management strategies when deploying ChatGPT for medical purposes. Furthermore, the tendency of these models to use disclaimers in conversations should be further investigated to understand the impact on patients\' experiences in practical applications.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    背景:大型语言模型显示出改善放射学工作流程的希望,但是它们在结构化放射任务(例如报告和数据系统(RADS)分类)上的表现仍未得到探索。
    目的:本研究旨在评估3个大型语言模型聊天机器人-Claude-2、GPT-3.5和GPT-4-在放射学报告中分配RADS类别并评估不同提示策略的影响。
    方法:这项横断面研究使用30个放射学报告(每个RADS标准10个)比较了3个聊天机器人,使用3级提示策略:零射,几枪,和指南PDF信息提示。这些病例的基础是2018年肝脏影像学报告和数据系统(LI-RADS),2022年肺部CT(计算机断层扫描)筛查报告和数据系统(Lung-RADS)和卵巢附件报告和数据系统(O-RADS)磁共振成像,由董事会认证的放射科医生精心准备。每份报告都进行了6次评估。两名失明的评论者评估了聊天机器人在患者级RADS分类和总体评级方面的反应。使用Fleissκ评估了跨重复的协议。
    结果:克劳德-2在总体评分中获得了最高的准确性,其中少量提示和指南PDF(提示-2),在6次运行中获得57%(17/30)的平均准确率,在k-pass投票中获得50%(15/30)的准确率。没有及时的工程,所有聊天机器人都表现不佳。结构化示例提示(prompt-1)的引入提高了所有聊天机器人整体评分的准确性。提供prompt-2进一步改进了Claude-2的性能,GPT-4未复制的增强。TheinterrunagreementwassubstantialforClaude-2(k=0.66foroverallratingandk=0.69forRADScategorization),对于GPT-4来说是公平的(两者的k=0.39),对于GPT-3.5来说是公平的(总体评分k=0.21,RADS分类k=0.39)。与Lung-RADS版本2022和O-RADS相比,2018年的所有聊天机器人均显示出更高的准确性(P<0.05);在2018年LI-RADS版本中,使用prompt-2,Claude-2实现了75%(45/60)的最高总体评分准确性。
    结论:当配备结构化提示和指导PDF时,Claude-2显示了根据既定标准(如LI-RADS版本2018)将RADS类别分配给放射学病例的潜力。然而,当前一代的聊天机器人滞后于根据最新的RADS标准对案件进行准确分类。
    BACKGROUND: Large language models show promise for improving radiology workflows, but their performance on structured radiological tasks such as Reporting and Data Systems (RADS) categorization remains unexplored.
    OBJECTIVE: This study aims to evaluate 3 large language model chatbots-Claude-2, GPT-3.5, and GPT-4-on assigning RADS categories to radiology reports and assess the impact of different prompting strategies.
    METHODS: This cross-sectional study compared 3 chatbots using 30 radiology reports (10 per RADS criteria), using a 3-level prompting strategy: zero-shot, few-shot, and guideline PDF-informed prompts. The cases were grounded in Liver Imaging Reporting & Data System (LI-RADS) version 2018, Lung CT (computed tomography) Screening Reporting & Data System (Lung-RADS) version 2022, and Ovarian-Adnexal Reporting & Data System (O-RADS) magnetic resonance imaging, meticulously prepared by board-certified radiologists. Each report underwent 6 assessments. Two blinded reviewers assessed the chatbots\' response at patient-level RADS categorization and overall ratings. The agreement across repetitions was assessed using Fleiss κ.
    RESULTS: Claude-2 achieved the highest accuracy in overall ratings with few-shot prompts and guideline PDFs (prompt-2), attaining 57% (17/30) average accuracy over 6 runs and 50% (15/30) accuracy with k-pass voting. Without prompt engineering, all chatbots performed poorly. The introduction of a structured exemplar prompt (prompt-1) increased the accuracy of overall ratings for all chatbots. Providing prompt-2 further improved Claude-2\'s performance, an enhancement not replicated by GPT-4. The interrun agreement was substantial for Claude-2 (k=0.66 for overall rating and k=0.69 for RADS categorization), fair for GPT-4 (k=0.39 for both), and fair for GPT-3.5 (k=0.21 for overall rating and k=0.39 for RADS categorization). All chatbots showed significantly higher accuracy with LI-RADS version 2018 than with Lung-RADS version 2022 and O-RADS (P<.05); with prompt-2, Claude-2 achieved the highest overall rating accuracy of 75% (45/60) in LI-RADS version 2018.
    CONCLUSIONS: When equipped with structured prompts and guideline PDFs, Claude-2 demonstrated potential in assigning RADS categories to radiology cases according to established criteria such as LI-RADS version 2018. However, the current generation of chatbots lags in accurately categorizing cases based on more recent RADS criteria.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    目的:本范围审查旨在审查其特征,应用程序,评估方法,以及在老年人中使用聊天机器人的挑战。
    方法:范围审查遵循Arksey和O\'Malley的方法框架,与Levac等人提出的修订。使用系统评价的首选报告项目和范围审查的荟萃分析扩展检查表报告研究结果。
    方法:所审查的文章主要关注老年人,在临床和非临床环境中进行的研究。
    方法:通过8个数据库检索了2010年1月至2023年5月发表的研究。本综述共确定并评估了29项研究。
    结果:结果表明,聊天机器人主要通过移动应用程序交付(n=11),他们中的大多数使用文本作为输入(n=16)和输出模式(n=13),其中大多数旨在改善老年人的整体福祉(n=9);大多数聊天机器人是为满足复杂的医疗保健需求(n=7)和健康信息收集(n=6)而设计的。本综述中捕获的聊天机器人的评估方法分为技术性能,用户可接受性,和有效性;将聊天机器人应用于老年人的挑战在于聊天机器人的设计,用户感知,和操作困难。
    结论:聊天机器人在老年人领域的使用仍在兴起,缺乏专门为老用户设计的选项。关于聊天机器人作为替代干预措施对健康影响的数据仍然有限。需要更标准化的评估标准和可靠的对照实验,以进一步研究聊天机器人在老年人中的有效性。
    OBJECTIVE: This scoping review aimed to review the characteristics, applications, evaluation approaches, and challenges regarding the use of chatbots in older adults.
    METHODS: The scoping review followed the methodological framework by Arksey and O\'Malley, with revisions proposed by Levac et al. The findings were reported using the Preferred Reporting Items for Systematic Review and Meta-Analysis Extension for Scoping Reviews checklist.
    METHODS: The reviewed articles primarily focused on older adults, with research conducted in both clinical and nonclinical settings.
    METHODS: Studies published from January 2010 to May 2023 were searched through 8 databases. A total of 29 studies were identified and evaluated in this review.
    RESULTS: Results showed that the chatbots were mainly delivered via mobile applications (n = 11), most of them used text as input (n = 16) and output modality (n = 13), and most of them targeted at improving the overall well-being of the older adults (n = 9); most chatbots were designed for fulfilling complex health care needs (n = 7) and health information collection (n = 6). Evaluation approaches of chatbots captured in this review were divided into technical performance, user acceptability, and effectiveness; challenges of applying chatbots to older adults lie in the design of the chatbot, user perception, and operational difficulties.
    CONCLUSIONS: The use of chatbots in the field of older adults is still emerging, with a lack of specifically designed options for older users. Data about the health impact of chatbots as alternative interventions were still limited. More standardized evaluation criteria and robust controlled experiments are needed for further research regarding the effectiveness of chatbots in older adults.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    聊天机器人在护理教育中的整合是一个快速发展的领域,具有潜在的变革性影响。这篇叙事综述旨在综合和分析护理教育中聊天机器人的现有文献。
    本研究旨在全面考察时间趋势,国际发行,研究设计,以及聊天机器人在护理教育中的意义。
    对3个数据库进行了全面搜索(PubMed,WebofScience,和Embase)遵循PRISMA(系统审查和荟萃分析的首选报告项目)流程图。
    共有40篇文章符合资格标准,2023年出版物显著增加(n=28,70%)。时间分析显示,从2021年到2023年,出版物显着激增,强调了日益增长的学术兴趣。地理上,台湾省做出了重大贡献(n=8,20%),其次是美国(n=6,15%)和韩国(n=4,10%)。研究设计多种多样,评论(n=8,20%)和社论(n=7,18%)占主导地位,展示了该领域研究的丰富性。
    将聊天机器人集成到护理教育中提出了一个有希望但相对未探索的途径。这篇综述强调了原创性研究的迫切需要,强调伦理考虑的重要性。
    UNASSIGNED: The integration of chatbots in nursing education is a rapidly evolving area with potential transformative impacts. This narrative review aims to synthesize and analyze the existing literature on chatbots in nursing education.
    UNASSIGNED: This study aims to comprehensively examine the temporal trends, international distribution, study designs, and implications of chatbots in nursing education.
    UNASSIGNED: A comprehensive search was conducted across 3 databases (PubMed, Web of Science, and Embase) following the PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) flow diagram.
    UNASSIGNED: A total of 40 articles met the eligibility criteria, with a notable increase of publications in 2023 (n=28, 70%). Temporal analysis revealed a notable surge in publications from 2021 to 2023, emphasizing the growing scholarly interest. Geographically, Taiwan province made substantial contributions (n=8, 20%), followed by the United States (n=6, 15%) and South Korea (n=4, 10%). Study designs varied, with reviews (n=8, 20%) and editorials (n=7, 18%) being predominant, showcasing the richness of research in this domain.
    UNASSIGNED: Integrating chatbots into nursing education presents a promising yet relatively unexplored avenue. This review highlights the urgent need for original research, emphasizing the importance of ethical considerations.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Editorial
    社交媒体的使用问题对个人的日常生活产生了许多负面影响,人际关系,身心健康,还有更多.目前,很少有方法和工具来缓解有问题的社交媒体,他们的潜力尚未充分发挥。新兴的大型语言模型(LLM)在为人们提供信息和帮助方面变得越来越流行,并被应用于生活的许多方面。在减轻有问题的社交媒体使用方面,ChatGPT等LLM可以通过充当用户的对话合作伙伴和网点来发挥积极作用,提供个性化的信息和资源,监控和干预有问题的社交媒体使用,还有更多.在这个过程中,我们应该认识到ChatGPT等LLM的巨大潜力和无限可能性,利用他们的优势更好地解决有问题的社交媒体使用问题,同时也承认ChatGPT技术的局限性和潜在的陷阱,如错误,问题解决的限制,隐私和安全问题,和潜在的过度依赖。当我们利用LLM的优势来解决社交媒体使用中的问题时,我们必须采取谨慎和道德的态度,警惕LLM在解决有问题的社交媒体使用方面可能产生的潜在不利影响,以更好地利用技术为个人和社会服务。
    The problematic use of social media has numerous negative impacts on individuals\' daily lives, interpersonal relationships, physical and mental health, and more. Currently, there are few methods and tools to alleviate problematic social media, and their potential is yet to be fully realized. Emerging large language models (LLMs) are becoming increasingly popular for providing information and assistance to people and are being applied in many aspects of life. In mitigating problematic social media use, LLMs such as ChatGPT can play a positive role by serving as conversational partners and outlets for users, providing personalized information and resources, monitoring and intervening in problematic social media use, and more. In this process, we should recognize both the enormous potential and endless possibilities of LLMs such as ChatGPT, leveraging their advantages to better address problematic social media use, while also acknowledging the limitations and potential pitfalls of ChatGPT technology, such as errors, limitations in issue resolution, privacy and security concerns, and potential overreliance. When we leverage the advantages of LLMs to address issues in social media usage, we must adopt a cautious and ethical approach, being vigilant of the potential adverse effects that LLMs may have in addressing problematic social media use to better harness technology to serve individuals and society.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Systematic Review
    背景:自2010年代中期以来,对话人工智能(AI;聊天机器人)在医疗保健中的使用已经显著扩大,特别是在COVID-19大流行期间卫生系统负担增加以及与医疗保健提供者面对面咨询受到限制的情况下。对话AI的一个新兴用途是捕捉不断发展的问题,并传达有关疫苗和疫苗接种的信息。
    目的:本系统评价的目的是检查有关对话AI在疫苗交流中的有效性的书面用途和证据。
    方法:本系统评价是按照PRISMA(系统评价和荟萃分析的首选报告项目)指南进行的。PubMed,WebofScience,PsycINFO,MEDLINE,Scopus,CINAHL完成,科克伦图书馆,Embase,认识论,全球卫生,全球指数Medicus,学术搜索完成,并在伦敦大学图书馆数据库中搜索了有关使用对话AI进行疫苗交流的论文。纳入标准是包括(1)用于疫苗交流的对话AI的记录实例和(2)对干预的影响和有效性的评估数据的研究。
    结果:删除重复项之后,审查确定了496条独特记录,然后根据标题和摘要进行筛选,其中38项被确定为全文审查。七项符合纳入标准,并在本综述的结果中进行了评估和总结。总的来说,迄今为止部署的疫苗聊天机器人的设计相对简单,主要用于向用户提供事实信息,以回答他们关于疫苗的问题。此外,聊天机器人已经被用于疫苗接种计划,约会提醒,揭穿错误信息,and,在某些情况下,用于疫苗咨询和说服。现有证据表明,聊天机器人可以对疫苗态度产生积极影响;然而,研究通常是探索性的,有些缺乏对照组或样本量很小。
    结论:该综述发现了对话AI对疫苗交流的潜在益处的证据。可能有助于疫苗聊天机器人有效性的因素包括它们实时提供可信和个性化信息的能力,聊天机器人平台的熟悉度和可访问性,以及与聊天机器人的互动对用户来说是“自然”的程度。然而,评估侧重于短期,聊天机器人对用户的直接影响。对话式AI的潜在长期和社会影响尚待分析。此外,现有的研究没有充分解决伦理学如何应用于围绕疫苗的对话人工智能领域。在可以预期疫苗通讯进一步数字化的背景下,所有这些领域都需要额外的高质量研究。
    Since the mid-2010s, use of conversational artificial intelligence (AI; chatbots) in health care has expanded significantly, especially in the context of increased burdens on health systems and restrictions on in-person consultations with health care providers during the COVID-19 pandemic. One emerging use for conversational AI is to capture evolving questions and communicate information about vaccines and vaccination.
    The objective of this systematic review was to examine documented uses and evidence on the effectiveness of conversational AI for vaccine communication.
    This systematic review was conducted following the PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) guidelines. PubMed, Web of Science, PsycINFO, MEDLINE, Scopus, CINAHL Complete, Cochrane Library, Embase, Epistemonikos, Global Health, Global Index Medicus, Academic Search Complete, and the University of London library database were searched for papers on the use of conversational AI for vaccine communication. The inclusion criteria were studies that included (1) documented instances of conversational AI being used for the purpose of vaccine communication and (2) evaluation data on the impact and effectiveness of the intervention.
    After duplicates were removed, the review identified 496 unique records, which were then screened by title and abstract, of which 38 were identified for full-text review. Seven fit the inclusion criteria and were assessed and summarized in the findings of this review. Overall, vaccine chatbots deployed to date have been relatively simple in their design and have mainly been used to provide factual information to users in response to their questions about vaccines. Additionally, chatbots have been used for vaccination scheduling, appointment reminders, debunking misinformation, and, in some cases, for vaccine counseling and persuasion. Available evidence suggests that chatbots can have a positive effect on vaccine attitudes; however, studies were typically exploratory in nature, and some lacked a control group or had very small sample sizes.
    The review found evidence of potential benefits from conversational AI for vaccine communication. Factors that may contribute to the effectiveness of vaccine chatbots include their ability to provide credible and personalized information in real time, the familiarity and accessibility of the chatbot platform, and the extent to which interactions with the chatbot feel \"natural\" to users. However, evaluations have focused on the short-term, direct effects of chatbots on their users. The potential longer-term and societal impacts of conversational AI have yet to be analyzed. In addition, existing studies do not adequately address how ethics apply in the field of conversational AI around vaccines. In a context where further digitalization of vaccine communication can be anticipated, additional high-quality research will be required across all these areas.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    对精神卫生支持的日益增长的需求凸显了对话代理人作为全球和中国人类支持者的重要性。这些药物可以增加可用性并降低心理健康支持的相对成本。提供的支持可以分为两种主要类型:认知和情感。关于该主题的现有工作主要集中在构建采用认知行为疗法(CBT)原则的代理上。这样的代理基于预定义的模板和练习来操作以提供认知支持。然而,关于使用此类代理的情感支持的研究是有限的。此外,大多数构建的代理人都用英语运作,强调在中国进行此类研究的重要性。为此,我们介绍Emohaa,通过CBT-Bot练习和引导对话提供认知支持的对话代理。它还通过ES-Bot在情感上支持用户,让他们发泄情绪问题.在这项研究中,我们分析了Emohaa在减轻精神困扰症状方面的有效性。
    按照RCT设计,本研究将参与者随机分为三组:Emohaa(CBT-Bot),Emohaa(Full),和控制。通过意向治疗(N=247)和PerProtocol(N=134)分析,结果表明,与对照组相比,使用两种类型的Emohaa的参与者在精神困扰症状方面经历了更显着的改善,包括抑郁症(F[2,244]=6.26,p=0.002),负面影响(F[2,244]=6.09,p=0.003),失眠(F[2,244]=3.69,p=0.026)。
    根据获得的结果和参与者对平台的满意度,我们得出结论,Emohaa是减少精神困扰的实用和有效的工具。
    UNASSIGNED: The growing demand for mental health support has highlighted the importance of conversational agents as human supporters worldwide and in China. These agents could increase availability and reduce the relative costs of mental health support. The provided support can be divided into two main types: cognitive and emotional. Existing work on this topic mainly focuses on constructing agents that adopt Cognitive Behavioral Therapy (CBT) principles. Such agents operate based on pre-defined templates and exercises to provide cognitive support. However, research on emotional support using such agents is limited. In addition, most of the constructed agents operate in English, highlighting the importance of conducting such studies in China. To this end, we introduce Emohaa, a conversational agent that provides cognitive support through CBT-Bot exercises and guided conversations. It also emotionally supports users through ES-Bot, enabling them to vent their emotional problems. In this study, we analyze the effectiveness of Emohaa in reducing symptoms of mental distress.
    UNASSIGNED: Following the RCT design, the current study randomly assigned participants into three groups: Emohaa (CBT-Bot), Emohaa (Full), and control. With both Intention-To-Treat (N=247) and PerProtocol (N=134) analyses, the results demonstrated that compared to the control group, participants who used two types of Emohaa experienced considerably more significant improvements in symptoms of mental distress, including depression (F[2,244]=6.26, p=0.002), negative affect (F[2,244]=6.09, p=0.003), and insomnia (F[2,244]=3.69, p=0.026).
    UNASSIGNED: Based on the obtained results and participants\' satisfaction with the platform, we concluded that Emohaa is a practical and effective tool for reducing mental distress.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Letter
    在过去的一个月里,一种新的人工智能模型,称为Chatbot生成预训练变换器(ChatGPT),由于它能够以人性化的方式处理和响应命令,因此在媒体和科学界受到了极大的关注。据报道,发射五天后,ChatGPT的注册用户数量超过一百万,两个月后,它的月活跃用户超过了1亿,使其成为历史上增长最快的消费者应用。ChatGPT的出现进一步带来了传染病领域的新思想和挑战。鉴于此,为了评估ChatGPT在传染病临床实践和科学研究中的潜在用途,我们使用公开的ChatGPT网页进行了简短的在线调查。此外,本研究还讨论了与该计划有关的相关社会和道德问题。
    Over the past month, a new AI model called Chatbot Generative Pre-trained Transformer (ChatGPT), has received enormous attention in the media and scientific communities due to its ability to process and respond to commands in a humanistic fashion. As reported, five days after its launch, the number of registered users of ChatGPT exceeded one million, and its monthly active users had exceeded 100 million two months later, making it the most rapidly growing consumer application in history. The advent of ChatGPT has further brought about new ideas and challenges in the realm of infectious disease. In view of this, in order to evaluate the potential use of ChatGPT in clinical practice and scientific research of infectious disease, we conducted a brief online survey by using the publicly available ChatGPT webpage. Also, the present study also talks about the relevant social and ethical issues related to this program.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Letter
    在大数据时代,生成人工智能(AI)模型目前正处于繁荣时期。Chatbot生成预训练变换器(ChatGPT),由OpenAI开发的大型语言模型(LLM)(旧金山,CA),是一种AI软件,可以根据收到的输入生成文本。在这项研究中,为了探索ChatGPT如何从AI维度对2022年突然爆发的Mpox给出反思和建议,我们小组与ChatGPT讨论了几个关于Mpox的问题。我们希望这次演讲可以从新的人工智能维度丰富我们对水痘的知识,并探索人类和人工智能并肩作战的可能性,以预防和遏制未来潜在的流行病或流行病。
    In the era of big data, generative artificial intelligence (AI) models are currently in a boom. The Chatbot Generative Pre-trained Transformer (ChatGPT), a large language model (LLM) developed by OpenAI (San Francisco, CA), is a type of AI software that could generate text based on the input it receives. In this study, in order to explore how ChatGPT could give reflections and suggestions about the sudden outbreak of Mpox in 2022 from the AI dimensions, our group talked with ChatGPT with several questions about Mpox. We hope this talk could enrich our knowledge on Mpox from the new AI dimensions and also explore the possibility of human and AI fight shoulder to shoulder for prevention and containment of the potential epidemics or pandemics in future.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

公众号