Chatbot

聊天机器人
  • 文章类型: Letter
    暂无摘要。
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    随着像ChatGPT这样的大型语言模型在各个行业中的应用越来越多,它在医疗领域的潜力,特别是在标准化考试中,已成为研究的重点。
    本研究的目的是评估ChatGPT的临床表现,重点关注其在中国国家医师资格考试(CNMLE)中的准确性和可靠性。
    CNMLE2022问题集,由500个单答案多选题组成,被重新分类为15个医学亚专科。从2023年4月24日至5月15日,每个问题在OpenAI平台上用中文进行了8到12次测试。考虑了三个关键因素:GPT-3.5和4.0版本,针对医疗亚专科定制的系统角色的提示指定,为了连贯性而重复。通过准确度阈值被建立为60%。采用χ2检验和κ值评估模型的准确性和一致性。
    GPT-4.0达到了72.7%的通过精度,显著高于GPT-3.5(54%;P<.001)。GPT-4.0重复反应的变异性低于GPT-3.5(9%vs19.5%;P<.001)。然而,两个模型都显示出相对较好的响应一致性,κ值分别为0.778和0.610。系统角色在数值上提高了GPT-4.0(0.3%-3.7%)和GPT-3.5(1.3%-4.5%)的准确性,并将变异性降低了1.7%和1.8%,分别(P>0.05)。在亚组分析中,ChatGPT在不同题型之间取得了相当的准确率(P>.05)。GPT-4.0在15个亚专业中的14个超过了准确性阈值,而GPT-3.5在第一次反应的15人中有7人这样做。
    GPT-4.0通过了CNMLE,并在准确性等关键领域优于GPT-3.5,一致性,和医学专科专业知识。添加系统角色不会显着增强模型的可靠性和答案的连贯性。GPT-4.0在医学教育和临床实践中显示出有希望的潜力,值得进一步研究。
    UNASSIGNED: With the increasing application of large language models like ChatGPT in various industries, its potential in the medical domain, especially in standardized examinations, has become a focal point of research.
    UNASSIGNED: The aim of this study is to assess the clinical performance of ChatGPT, focusing on its accuracy and reliability in the Chinese National Medical Licensing Examination (CNMLE).
    UNASSIGNED: The CNMLE 2022 question set, consisting of 500 single-answer multiple choices questions, were reclassified into 15 medical subspecialties. Each question was tested 8 to 12 times in Chinese on the OpenAI platform from April 24 to May 15, 2023. Three key factors were considered: the version of GPT-3.5 and 4.0, the prompt\'s designation of system roles tailored to medical subspecialties, and repetition for coherence. A passing accuracy threshold was established as 60%. The χ2 tests and κ values were employed to evaluate the model\'s accuracy and consistency.
    UNASSIGNED: GPT-4.0 achieved a passing accuracy of 72.7%, which was significantly higher than that of GPT-3.5 (54%; P<.001). The variability rate of repeated responses from GPT-4.0 was lower than that of GPT-3.5 (9% vs 19.5%; P<.001). However, both models showed relatively good response coherence, with κ values of 0.778 and 0.610, respectively. System roles numerically increased accuracy for both GPT-4.0 (0.3%-3.7%) and GPT-3.5 (1.3%-4.5%), and reduced variability by 1.7% and 1.8%, respectively (P>.05). In subgroup analysis, ChatGPT achieved comparable accuracy among different question types (P>.05). GPT-4.0 surpassed the accuracy threshold in 14 of 15 subspecialties, while GPT-3.5 did so in 7 of 15 on the first response.
    UNASSIGNED: GPT-4.0 passed the CNMLE and outperformed GPT-3.5 in key areas such as accuracy, consistency, and medical subspecialty expertise. Adding a system role insignificantly enhanced the model\'s reliability and answer coherence. GPT-4.0 showed promising potential in medical education and clinical practice, meriting further study.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Editorial
    传统的人工智能(AI)工具已经在临床放射学中实现,用于病变检测和决策。生成人工智能(GenAI),相比之下,是机器学习的一个新子集,它基于数据概率来创建内容,提供了许多能力,但也有不确定性。多学科合作对于安全利用GenAI的力量至关重要,因为它改变了医学。本文建议在放射性社会中建立一个GenAI工作组,包括美国放射学院(ACR),医学成像信息学学会(SIIM),北美放射学会(RSNA)欧洲放射学会(ESR),大学放射科医师协会(AUR),和美国伦琴射线学会(ARRS)将其融入临床护理,卫生政策,和教育。在本文中,我们探讨了具有指南的工作组将如何帮助放射科医师和受训者制定基本策略,将不断发展的AI相关技术整合到临床实践中。
    Traditional artificial intelligence (AI) tools have already been implemented in clinical radiology for lesion detection and decision-making. Generative AI (GenAI), comparingly, is a new subset of machine learning that functions based on data probabilities to create content, offering numerous capabilities yet also uncertainties. Multidisciplinary collaboration is essential in safely harnessing the power of GenAI as it transforms medicine. This paper proposes creating a GenAI task force among radiological societies, including the American College of Radiology (ACR), Society of Imaging Informatics in Medicine (SIIM), Radiological Society of North America (RSNA), European Society of Radiology (ESR), Association of University Radiologists (AUR), and American Roentgen Ray Society (ARRS) for its integration into clinical care, health policy, and education. In this paper, we explore how a task force with guidelines will help radiologists and trainees develop essential strategies for integrating evolving AI-related technologies into clinical practice.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    背景:HIV暴露前预防(PrEP)是预防顺性女性之间HIV传播的重要生物医学策略。尽管其有效性已被证明,在整个PrEP护理连续过程中,黑人女性的比例仍然严重不足,面临障碍,如获得护理的机会有限,医学上的不信任,以及交叉的种族或艾滋病毒耻辱。解决这些差异对于改善该社区的艾滋病毒预防成果至关重要。另一方面,护士从业人员(NPs)在PrEP利用中起着关键作用,但由于缺乏意识,代表性不足,缺乏人力资源,支持不足。配备人工智能(AI)和先进的大型语言模型的快速发展,聊天机器人有效地促进了医疗交流和与各个领域的医疗联系,包括艾滋病毒预防和PrEP护理。
    目的:我们的研究通过自然语言处理算法利用NPs的整体护理能力和AI的力量,提供有针对性的,以患者为中心促进PrEP护理。我们的首要目标是创建一个护士主导的,利益相关者包容性,和人工智能驱动的计划,以促进顺性黑人女性的PrEP利用,最终分三个阶段加强这一弱势群体的艾滋病毒预防工作。该项目旨在缓解健康差距,推进创新,基于技术的解决方案。
    方法:该研究使用混合方法设计,涉及与关键利益相关者的半结构化访谈,包括50名符合PrEP资格的黑人女性,10个NP,以及代表各种社会经济背景的社区顾问委员会。AI驱动的聊天机器人使用HumanX技术和SmartBot360的健康保险可移植性和责任法案兼容框架开发,以确保数据隐私和安全。这项研究历时18个月,包括3个阶段:探索,发展,和评价。
    结果:截至2024年5月,第一阶段的机构审查委员会方案已获得批准。我们计划在2024年9月开始招募黑人女性和NP,目的是收集信息以了解他们对聊天机器人开发的偏好。虽然机构审查委员会对第二阶段和第三阶段的批准仍在进行中,我们在参与者招募网络方面取得了重大进展。我们计划很快进行数据收集,随着研究的进展,将提供招聘和数据收集进展的进一步更新。
    结论:AI驱动的聊天机器人提供了一种新颖的方法来改善黑人女性的PrEP护理利用率,有机会减少护理障碍,并促进无污名化的环境。然而,卫生公平和数字鸿沟方面的挑战仍然存在,强调需要有文化能力的设计和强大的数据隐私协议。这项研究的意义超出了PrEP护理,提出了一个可扩展的模型,可以解决更广泛的健康差距。
    PRR1-10.2196/59975。
    BACKGROUND: HIV pre-exposure prophylaxis (PrEP) is a critical biomedical strategy to prevent HIV transmission among cisgender women. Despite its proven effectiveness, Black cisgender women remain significantly underrepresented throughout the PrEP care continuum, facing barriers such as limited access to care, medical mistrust, and intersectional racial or HIV stigma. Addressing these disparities is vital to improving HIV prevention outcomes within this community. On the other hand, nurse practitioners (NPs) play a pivotal role in PrEP utilization but are underrepresented due to a lack of awareness, a lack of human resources, and insufficient support. Equipped with the rapid evolution of artificial intelligence (AI) and advanced large language models, chatbots effectively facilitate health care communication and linkage to care in various domains, including HIV prevention and PrEP care.
    OBJECTIVE: Our study harnesses NPs\' holistic care capabilities and the power of AI through natural language processing algorithms, providing targeted, patient-centered facilitation for PrEP care. Our overarching goal is to create a nurse-led, stakeholder-inclusive, and AI-powered program to facilitate PrEP utilization among Black cisgender women, ultimately enhancing HIV prevention efforts in this vulnerable group in 3 phases. This project aims to mitigate health disparities and advance innovative, technology-based solutions.
    METHODS: The study uses a mixed methods design involving semistructured interviews with key stakeholders, including 50 PrEP-eligible Black women, 10 NPs, and a community advisory board representing various socioeconomic backgrounds. The AI-powered chatbot is developed using HumanX technology and SmartBot360\'s Health Insurance Portability and Accountability Act-compliant framework to ensure data privacy and security. The study spans 18 months and consists of 3 phases: exploration, development, and evaluation.
    RESULTS: As of May 2024, the institutional review board protocol for phase 1 has been approved. We plan to start recruitment for Black cisgender women and NPs in September 2024, with the aim to collect information to understand their preferences regarding chatbot development. While institutional review board approval for phases 2 and 3 is still in progress, we have made significant strides in networking for participant recruitment. We plan to conduct data collection soon, and further updates on the recruitment and data collection progress will be provided as the study advances.
    CONCLUSIONS: The AI-powered chatbot offers a novel approach to improving PrEP care utilization among Black cisgender women, with opportunities to reduce barriers to care and facilitate a stigma-free environment. However, challenges remain regarding health equity and the digital divide, emphasizing the need for culturally competent design and robust data privacy protocols. The implications of this study extend beyond PrEP care, presenting a scalable model that can address broader health disparities.
    UNASSIGNED: PRR1-10.2196/59975.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    患者在临床治疗前通过多种信息渠道获取有关其骨科手术资源的相关信息。最近,人工智能(AI)驱动的聊天机器人已经成为患者的另一个信息来源。当前开发的AI聊天技术ChatGPT(OpenAILP)是用于此类目的的应用程序,并且已迅速普及,包括患者教育。这项研究旨在评估ChatGPT是否可以正确回答有关假体周围感染(PJI)的常见问题(FAQ)。
    在15个国际临床专家中心的网站上发现了12个关于髋关节和膝关节置换术后PJI的常见问题。ChatGPT面临这些问题,一个多学科团队使用基于证据的方法分析了其回答的准确性。反应分为四组:(1)不需要额外改善的出色反应;(2)需要少量改善的满意反应;(3)需要适度改善的满意反应;(4)需要大量改善的不满意反应。
    通过对聊天机器人给出的响应的分析,没有答复收到“不满意”评级;一个不需要任何更正;大多数答复要求低(12个中的7个)或中等(12个中的4个)澄清。尽管一些答复需要最少的澄清,聊天机器人的反应通常是公正的,以证据为基础的,即使被问到有争议的问题。
    AI聊天机器人ChatGPT能够有效地回答寻求PJI诊断和治疗信息的患者的常见问题。给定的信息也以可以被认为是患者可理解的方式编写。聊天机器人可能是未来患者教育和理解PJI治疗的宝贵临床工具。进一步的研究应评估其使用和接受PJI患者。
    UNASSIGNED: Patients access relevant information concerning their orthopaedic surgery resources through multiple information channels before presenting for clinical treatment. Recently, artificial intelligence (AI)-powered chatbots have become another source of information for patients. The currently developed AI chat technology ChatGPT (OpenAI LP) is an application for such purposes and it has been rapidly gaining popularity, including for patient education. This study sought to evaluate whether ChatGPT can correctly answer frequently asked questions (FAQ) regarding periprosthetic joint infection (PJI).
    UNASSIGNED: Twelve FAQs about PJI after hip and knee arthroplasty were identified from the websites of fifteen international clinical expert centres. ChatGPT was confronted with these questions and its responses were analysed for their accuracy using an evidence-based approach by a multidisciplinary team. Responses were categorised in four groups: (1) Excellent response that did not require additional improvement; (2) Satisfactory responses that required a small amount of improvement; (3) Satisfactory responses that required moderate improvement; and (4) Unsatisfactory responses that required a large amount of improvement.
    UNASSIGNED: From the analysis of the responses given by the chatbot, no reply received an \'unsatisfactory\' rating; one did not require any correction; and the majority of the responses required low (7 out of 12) or moderate (4 out of 12) clarification. Although a few responses required minimal clarification, the chatbot responses were generally unbiased and evidence-based, even when asked controversial questions.
    UNASSIGNED: The AI-chatbot ChatGPT was able to effectively answer the FAQs of patients seeking information around PJI diagnosis and treatment. The given information was also written in a manner that can be assumed to be understandable by patients. The chatbot could be a valuable clinical tool for patient education and understanding around PJI treatment in the future. Further studies should evaluate its use and acceptance by patients with PJI.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    为了评估响应能力,在公共医疗系统耳鼻喉科工作竞争考试中,ChatGPT3.5和互联网连接的GPT-4引擎(MicrosoftCopilot),以耳鼻喉科专家的真实分数为对照组。2023年9月,将135个分为理论和实践部分的问题输入到ChatGPT3.5和连接互联网的GPT-4中。将AI反应的准确性与参加考试的耳鼻喉科医生的官方结果进行了比较,采用Stata14.2进行统计分析。副驾驶(GPT-4)的表现优于ChatGPT3.5。副驾驶取得88.5分的成绩,而ChatGPT得了60分。两个AI的错误答案都存在差异。尽管ChatGPT很熟练,Copilot表现出卓越的性能,在参加考试的108名耳鼻喉科医生中排名第二,而ChatGPT排在第83位。与ChatGPT3.5相比,由具有互联网访问功能的GPT-4(Copilot)提供的聊天在回答多项选择的医疗问题方面表现出卓越的性能。
    To evaluate the response capabilities, in a public healthcare system otolaryngology job competition examination, of ChatGPT 3.5 and an internet-connected GPT-4 engine (Microsoft Copilot) with the real scores of otolaryngology specialists as the control group. In September 2023, 135 questions divided into theoretical and practical parts were input into ChatGPT 3.5 and an internet-connected GPT-4. The accuracy of AI responses was compared with the official results from otolaryngologists who took the exam, and statistical analysis was conducted using Stata 14.2. Copilot (GPT-4) outperformed ChatGPT 3.5. Copilot achieved a score of 88.5 points, while ChatGPT scored 60 points. Both AIs had discrepancies in their incorrect answers. Despite ChatGPT\'s proficiency, Copilot displayed superior performance, ranking as the second-best score among the 108 otolaryngologists who took the exam, while ChatGPT was placed 83rd. A chat powered by GPT-4 with internet access (Copilot) demonstrates superior performance in responding to multiple-choice medical questions compared to ChatGPT 3.5.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    人工智能(AI)驱动的心理健康会话代理的日益普及,需要全面了解用户参与度和用户对该技术的看法。本研究旨在通过关注Wysa来填补现有的知识空白,一种商业上可用的移动会话代理,旨在提供个性化的心理健康支持。
    1月之间共发布了159条用户评论,2020年3月,2024年,在Wysa应用程序的GooglePlay页面上进行了收集。然后使用主题分析对收集的数据进行开放和归纳编码。
    用户评论中出现了七个主要主题:“信任环境促进福祉”,\“无处不在的访问提供实时支持\”,“AI限制会降低用户体验”,“Wysa的感知有效性”,“渴望有凝聚力和可预测的互动”,“人工智能中的人性受到欢迎”,和“需要改进用户界面”。这些主题突出了人工智能驱动的心理健康对话代理的好处和局限性。
    用户发现Wysa有效地培养了与用户的牢固联系,鼓励他们参与应用程序,并采取积极的步骤,对情绪弹性和自我完善。然而,它的AI需要进行一些改进,以增强应用程序的用户体验。这些发现有助于设计和实施更有效的,伦理,和用户一致的AI驱动的心理健康支持系统。
    UNASSIGNED: The increasing prevalence of artificial intelligence (AI)-driven mental health conversational agents necessitates a comprehensive understanding of user engagement and user perceptions of this technology. This study aims to fill the existing knowledge gap by focusing on Wysa, a commercially available mobile conversational agent designed to provide personalized mental health support.
    UNASSIGNED: A total of 159 user reviews posted between January, 2020 and March, 2024, on the Wysa app\'s Google Play page were collected. Thematic analysis was then used to perform open and inductive coding of the collected data.
    UNASSIGNED: Seven major themes emerged from the user reviews: \"a trusting environment promotes wellbeing\", \"ubiquitous access offers real-time support\", \"AI limitations detract from the user experience\", \"perceived effectiveness of Wysa\", \"desire for cohesive and predictable interactions\", \"humanness in AI is welcomed\", and \"the need for improvements in the user interface\". These themes highlight both the benefits and limitations of the AI-driven mental health conversational agents.
    UNASSIGNED: Users find that Wysa is effective in fostering a strong connection with its users, encouraging them to engage with the app and take positive steps towards emotional resilience and self-improvement. However, its AI needs several improvements to enhance user experience with the application. The findings contribute to the design and implementation of more effective, ethical, and user-aligned AI-driven mental health support systems.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    背景:心脏代谢疾病(CMD)是一组相互关联的疾病,包括心力衰竭和糖尿病,增加心血管和代谢并发症的风险。拥有CMD的澳大利亚人数量不断增加,因此需要为管理这些条件的人制定新的策略,例如数字健康干预。数字健康干预措施在支持CMD人群方面的有效性取决于用户使用工具的程度。使用对话代理加强数字健康干预,使用自然语言与人互动的技术,可能会因为它们类似人类的属性而增强参与度。迄今为止,没有系统评价收集有关设计特征如何影响支持CMD患者的对话式代理干预的参与的证据.这项审查旨在解决这一差距,从而指导开发人员为CMD管理创建更具吸引力和有效的工具。
    目的:本系统评价的目的是综合有关对话代理干预设计特征及其对管理CMD的人员参与的影响的证据。
    方法:审查是根据Cochrane干预措施系统审查手册进行的,并根据PRISMA(系统审查和荟萃分析的首选报告项目)指南进行报告。搜索将在Ovid(Medline)进行,WebofScience,和Scopus数据库,它将在提交手稿之前再次运行。纳入标准将包括主要研究研究报告对话代理启用的干预措施,包括接触措施,成人CMD数据提取将寻求捕获CMD人群对使用对话代理干预的观点。JoannaBriggs研究所的关键评估工具将用于评估收集的证据的整体质量。
    结果:该评论于2023年5月启动,并于2023年6月在国际前瞻性系统评论注册中心(PROSPERO)注册,然后进行标题和摘要筛选。论文全文筛选已于2023年7月完成,数据提取于2023年8月开始。最终搜索于2024年4月进行,然后最终完成审查,手稿于2024年7月提交同行评审。
    结论:本综述将综合与对话代理启用的干预设计特征及其对CMD人群参与的影响有关的各种观察结果。这些观察结果可用于指导开发更具吸引力的对话代理干预措施,从而增加了定期使用干预措施的可能性,并改善了CMD健康结果。此外,这篇综述将确定文献中关于参与度如何报告的差距,从而突出了未来探索的领域,并支持研究人员推进对会话代理启用的干预措施的理解。
    背景:PROSPEROCRD42023431579;https://tinyurl.com/55cxkm26。
    DERR1-10.2196/52973。
    BACKGROUND: Cardiometabolic diseases (CMDs) are a group of interrelated conditions, including heart failure and diabetes, that increase the risk of cardiovascular and metabolic complications. The rising number of Australians with CMDs has necessitated new strategies for those managing these conditions, such as digital health interventions. The effectiveness of digital health interventions in supporting people with CMDs is dependent on the extent to which users engage with the tools. Augmenting digital health interventions with conversational agents, technologies that interact with people using natural language, may enhance engagement because of their human-like attributes. To date, no systematic review has compiled evidence on how design features influence the engagement of conversational agent-enabled interventions supporting people with CMDs. This review seeks to address this gap, thereby guiding developers in creating more engaging and effective tools for CMD management.
    OBJECTIVE: The aim of this systematic review is to synthesize evidence pertaining to conversational agent-enabled intervention design features and their impacts on the engagement of people managing CMD.
    METHODS: The review is conducted in accordance with the Cochrane Handbook for Systematic Reviews of Interventions and reported in accordance with PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) guidelines. Searches will be conducted in the Ovid (Medline), Web of Science, and Scopus databases, which will be run again prior to manuscript submission. Inclusion criteria will consist of primary research studies reporting on conversational agent-enabled interventions, including measures of engagement, in adults with CMD. Data extraction will seek to capture the perspectives of people with CMD on the use of conversational agent-enabled interventions. Joanna Briggs Institute critical appraisal tools will be used to evaluate the overall quality of evidence collected.
    RESULTS: This review was initiated in May 2023 and was registered with the International Prospective Register of Systematic Reviews (PROSPERO) in June 2023, prior to title and abstract screening. Full-text screening of articles was completed in July 2023 and data extraction began August 2023. Final searches were conducted in April 2024 prior to finalizing the review and the manuscript was submitted for peer review in July 2024.
    CONCLUSIONS: This review will synthesize diverse observations pertaining to conversational agent-enabled intervention design features and their impacts on engagement among people with CMDs. These observations can be used to guide the development of more engaging conversational agent-enabled interventions, thereby increasing the likelihood of regular intervention use and improved CMD health outcomes. Additionally, this review will identify gaps in the literature in terms of how engagement is reported, thereby highlighting areas for future exploration and supporting researchers in advancing the understanding of conversational agent-enabled interventions.
    BACKGROUND: PROSPERO CRD42023431579; https://tinyurl.com/55cxkm26.
    UNASSIGNED: DERR1-10.2196/52973.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    背景:聊天机器人的普及,特别是OpenAI在公众中的ChatGPT及其在医疗保健领域的实用性是当前争议的话题。当前的研究旨在评估ChatGPT对父母提出的询问的回答的可靠性和准确性,特别关注一系列儿科眼科和斜视疾病。
    方法:通过主题分析收集患者查询,并将其提交给ChatGPT3.5版本,每个案例有3个独特的实例。这些问题分为12个领域,共817个独特问题。所有反应均由两名经验丰富的儿科眼科医生以Likert量表格式对反应质量进行评分。使用Flesch-Kincaid等级(FKGL)和字符计数评估所有问题的可读性。
    结果:共638个(78.09%)问题得分完全正确,156(19.09%)得分正确,但不完整,只有23(2.81%)得分部分不正确。没有一个回答被认为是完全不正确的。平均FKGL评分为14.49[95%CI14.4004-14.5854],平均字符计数为1825.33[95CI1791.95-1858.7],p=0.831和0.697。最小和最大FKGL评分分别为10.6和18.34。FKGL预测字符计数,R²=.012,F(1815)=10.26,p=.001。
    结论:ChatGPT为大多数问题提供了准确可靠的信息。问题的可读性远高于成年人通常要求的标准,这是关于。尽管有这些限制,显然,这项技术将在医疗保健行业发挥重要作用。
    BACKGROUND: The rise in popularity of chatbots, particularly ChatGPT by OpenAI among the general public and its utility in the healthcare field is a topic of present controversy. The current study aimed at assessing the reliability and accuracy of ChatGPT\'s responses to inquiries posed by parents, specifically focusing on a range of pediatric ophthalmological and strabismus conditions.
    METHODS: Patient queries were collected via a thematic analysis and posed to ChatGPT 3.5 version across 3 unique instances each. The questions were divided into 12 domains totalling 817 unique questions. All responses were scored on the response quality by two experienced pediatric ophthalmologists in a Likert-scale format. All questions were evaluated for readability using the Flesch-Kincaid Grade Level (FKGL) and character counts.
    RESULTS: A total of 638 (78.09%) questions were scored to be perfectly correct, 156 (19.09%) were scored correct but incomplete and only 23 (2.81%) were scored to be partially incorrect. None of the responses were scored to be completely incorrect. Average FKGL score was 14.49 [95% CI 14.4004-14.5854] and the average character count was 1825.33 [95%CI 1791.95-1858.7] with p = 0.831 and 0.697 respectively. The minimum and maximum FKGL scores were 10.6 and 18.34 respectively. FKGL predicted character count, R²=.012, F(1,815) = 10.26, p = .001.
    CONCLUSIONS: ChatGPT provided accurate and reliable information for a majority of the questions. The readability of the questions was much above the typically required standards for adults, which is concerning. Despite these limitations, it is evident that this technology will play a significant role in the healthcare industry.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    背景:吸烟构成重大的公共卫生风险。聊天机器人可以作为一种可访问和有用的工具来促进戒烟,因为它们具有很高的可访问性和促进长期个性化交互的潜力。为了提高有效性和可接受性,仍然需要确定和评估这些聊天机器人的咨询策略,这在以前的研究中没有得到全面解决。
    目的:本研究旨在为此类聊天机器人确定有效的咨询策略,以支持戒烟。此外,我们试图深入了解吸烟者对聊天机器人的期望和体验。
    方法:这项混合方法研究结合了基于网络的实验和半结构化访谈。吸烟者(N=229)与动机性访谈(MI)风格(n=112,48.9%)或对抗性咨询风格(n=117,51.1%)的聊天机器人进行了互动。两者都与停止有关(即,退出意向和自我效能感)和与用户体验相关的结果(即,订婚,治疗联盟,感知到的同理心,和互动满意度)进行评估。对16名参与者进行了半结构化访谈,8(50%)从每个条件,并采用专题分析法对数据进行分析。
    结果:多变量方差分析的结果表明,参与者对MI(与对抗性咨询)聊天机器人的总体评分明显更高。后续判别分析显示,MI聊天机器人的更好感知主要由用户体验相关结果解释,与戒烟相关的结果起的作用较小。探索性分析表明,两种情况下的吸烟者在聊天机器人互动后都报告了戒烟意愿和自我效能感的增加。访谈结果说明了几种结构(例如,情感态度和参与度)解释人们以前的期望以及与聊天机器人的及时和回顾性的经验。
    结论:结果证实,聊天机器人是激励戒烟的有前途的工具,使用MI可以改善用户体验。我们没有找到对MI激励戒烟的额外支持,并讨论了可能的原因。吸烟者在戒烟过程中既表达了关系需求,也表达了工具需求。讨论了对未来研究和实践的启示。
    BACKGROUND: Cigarette smoking poses a major public health risk. Chatbots may serve as an accessible and useful tool to promote cessation due to their high accessibility and potential in facilitating long-term personalized interactions. To increase effectiveness and acceptability, there remains a need to identify and evaluate counseling strategies for these chatbots, an aspect that has not been comprehensively addressed in previous research.
    OBJECTIVE: This study aims to identify effective counseling strategies for such chatbots to support smoking cessation. In addition, we sought to gain insights into smokers\' expectations of and experiences with the chatbot.
    METHODS: This mixed methods study incorporated a web-based experiment and semistructured interviews. Smokers (N=229) interacted with either a motivational interviewing (MI)-style (n=112, 48.9%) or a confrontational counseling-style (n=117, 51.1%) chatbot. Both cessation-related (ie, intention to quit and self-efficacy) and user experience-related outcomes (ie, engagement, therapeutic alliance, perceived empathy, and interaction satisfaction) were assessed. Semistructured interviews were conducted with 16 participants, 8 (50%) from each condition, and data were analyzed using thematic analysis.
    RESULTS: Results from a multivariate ANOVA showed that participants had a significantly higher overall rating for the MI (vs confrontational counseling) chatbot. Follow-up discriminant analysis revealed that the better perception of the MI chatbot was mostly explained by the user experience-related outcomes, with cessation-related outcomes playing a lesser role. Exploratory analyses indicated that smokers in both conditions reported increased intention to quit and self-efficacy after the chatbot interaction. Interview findings illustrated several constructs (eg, affective attitude and engagement) explaining people\'s previous expectations and timely and retrospective experience with the chatbot.
    CONCLUSIONS: The results confirmed that chatbots are a promising tool in motivating smoking cessation and the use of MI can improve user experience. We did not find extra support for MI to motivate cessation and have discussed possible reasons. Smokers expressed both relational and instrumental needs in the quitting process. Implications for future research and practice are discussed.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

公众号