Human–computer interaction

人机交互
  • 文章类型: Journal Article
    放射学短缺影响了全球一半以上的人口,人工智能彻底改变医疗诊断和治疗的潜力变得越来越重要。然而,缺乏医疗专业人员的信任阻碍了AI模型在健康科学中的广泛采用。可解释AI(XAI)旨在通过识别偏见并提供透明的解释来增加对黑盒模型的信任和理解。这是从医学领域的角度探讨可解释用户界面(XUI)的第一个调查,分析当前医疗XAI系统中采用的可视化和交互方式。我们按照PRISMA方法分析了42个可解释的接口,强调作为解释过程的一部分,有效地向用户传达信息的关键作用。我们贡献了界面设计属性的分类法,并确定了五个不同的研究论文集群。未来的研究方向包括医疗决策支持中的可竞争性,对图像的反事实解释,并利用大型语言模型来增强医疗保健中的XAI接口。
    With radiology shortages affecting over half of the global population, the potential of artificial intelligence to revolutionize medical diagnosis and treatment is ever more important. However, lacking trust from medical professionals hinders the widespread adoption of AI models in health sciences. Explainable AI (XAI) aims to increase trust and understanding of black box models by identifying biases and providing transparent explanations. This is the first survey that explores explainable user interfaces (XUI) from a medical domain perspective, analysing the visualization and interaction methods employed in current medical XAI systems. We analysed 42 explainable interfaces following the PRISMA methodology, emphasizing the critical role of effectively conveying information to users as part of the explanation process. We contribute a taxonomy of interface design properties and identify five distinct clusters of research papers. Future research directions include contestability in medical decision support, counterfactual explanations for images, and leveraging Large Language Models to enhance XAI interfaces in healthcare.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    本研究的目的是开发一种实用的软件解决方案,用于使用两个手臂实时识别手语单词。这将促进听力受损者与能听见者之间的交流。我们知道使用不同技术开发的几种手语识别系统,包括摄像头,臂章,和手套。然而,我们在这项研究中提出的系统以其实用性而脱颖而出,利用两臂的表面肌电图(肌肉活动)和惯性测量单元(运动动力学)数据。我们解决了其他方法的缺点,比如高成本,由于环境光和障碍物的低精度,和复杂的硬件要求,这限制了它们的实际应用。我们的软件可以使用本研究特有的数字信号处理和机器学习方法在不同的操作系统上运行。对于测试,我们根据其在日常生活中的使用频率创建了一个包含80个单词的数据集,并进行了彻底的特征提取过程。我们使用各种分类器和参数测试了识别性能,并比较了结果。随机森林算法是最成功的,达到惊人的99.875%的准确度,而朴素贝叶斯算法的成功率最低,准确率为87.625%。新系统有望显着改善听力障碍者的沟通,并确保无缝集成到日常生活中,而不会影响用户的舒适度或生活质量。
    The aim of this study is to develop a practical software solution for real-time recognition of sign language words using two arms. This will facilitate communication between hearing-impaired individuals and those who can hear. We are aware of several sign language recognition systems developed using different technologies, including cameras, armbands, and gloves. However, the system we propose in this study stands out for its practicality, utilizing surface electromyography (muscle activity) and inertial measurement unit (motion dynamics) data from both arms. We address the drawbacks of other methods, such as high costs, low accuracy due to ambient light and obstacles, and complex hardware requirements, which have limited their practical application. Our software can run on different operating systems using digital signal processing and machine learning methods specific to this study. For the test, we created a dataset of 80 words based on their frequency of use in daily life and performed a thorough feature extraction process. We tested the recognition performance using various classifiers and parameters and compared the results. The random forest algorithm emerged as the most successful, achieving a remarkable 99.875% accuracy, while the naïve Bayes algorithm had the lowest success rate with 87.625% accuracy. The new system promises to significantly improve communication for people with hearing disabilities and ensures seamless integration into daily life without compromising user comfort or lifestyle quality.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    目的:本研究探讨了增强现实(AR)在内镜经蝶入路(TSA)中的潜在应用。虽然先前的研究已经解决了TSAAR中的技术挑战,本文从以人为本的设计角度探讨了设计因素如何改善神经外科医生的AR。
    方法:初步定性研究包括对TSA程序的观察(n=2)和对神经外科医生的半结构化访谈(n=4)。这些为AR模型的设计提供了信息,与神经外科医生进行了评估(n=3)。然后在Unity3D中开发了一个交互式低保真原型-“AR辅助的跨序列方法导航(ANTSA)”。一项用户研究(n=4)通过上下文访谈评估了ANTSA的低保真度原型,提供有关设计因素的反馈。
    结果:AR可视化可能有助于简化鞍相并减少术中错误,例如过度或不充分的暴露。主要设计建议包括精益网格渲染,直观的调色板,和可选的结构突出显示。
    结论:这项研究提出了以用户为中心的设计指南,以改善TSA鞍相的感觉形成和手术工作流程,以改善临床结果为目标。讨论了AR可以为工作流程带来的具体改进,以及外科医生的保留及其在培训经验不足的医生方面的可能应用。
    OBJECTIVE: This study investigates the potential utility of augmented reality (AR) in the endoscopic transsphenoidal approach (TSA). While previous research has addressed technical challenges in AR for TSA, this paper explores how design factors can improve AR for neurosurgeons from a human-centred design perspective.
    METHODS: Preliminary qualitative research involved observations of TSA procedures ( n = 2 ) and semi-structured interviews with neurosurgeons ( n = 4 ). These informed the design of an AR mockup, which was evaluated with neurosurgeons ( n = 3 ). An interactive low-fidelity prototype-the \"AR-assisted Navigation for the TransSphenoidal Approach (ANTSA)\"-was then developed in Unity 3D. A user study ( n = 4 ) evaluated the low-fidelity prototype of ANTSA through contextual interviews, providing feedback on design factors.
    RESULTS: AR visualisations may be beneficial in streamlining the sellar phase and reducing intraoperative errors such as excessive or inadequate exposure. Key design recommendations include a lean mesh rendering, an intuitive colour palette, and optional structure highlighting.
    CONCLUSIONS: This research presents user-centred design guidelines to improve sensemaking and surgical workflow in the sellar phase of TSA, with the goal of improving clinical outcomes. The specific improvements that AR could bring to the workflow are discussed along with surgeons\' reservations and its possible application towards training less experienced physicians.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    聊天机器人可以影响大规模的行为改变,因为它们可以通过社交媒体访问。灵活,可扩展,并自动收集数据。然而,关于聊天机器人管理的行为改变干预措施的可行性和有效性的研究很少。在聊天机器人中实施既定的行为改变干预措施的有效性得不到保证,鉴于独特的人机交互动力学。我们通过信息提供和嵌入式动画对基于聊天机器人的行为改变进行了试点测试。我们评估了聊天机器人是否可以在大流行期间增加理解和采取保护性行为的意图。59名文化和语言不同的参与者接受了同情干预,指数增长干预,或者不干预。我们测量了参与者的COVID-19测试意图,并测量了他们在聊天机器人互动前后的待在家里的态度。我们发现保护行为的不确定性降低。指数增长干预增加了参与者的测试意图。这项研究提供了初步证据,表明聊天机器人可以引发行为改变,在多元化和代表性不足的群体中应用。
    Chatbots can effect large-scale behaviour change because they are accessible through social media, flexible, scalable, and gather data automatically. Yet research on the feasibility and effectiveness of chatbot-administered behaviour change interventions is sparse. The effectiveness of established behaviour change interventions when implemented in chatbots is not guaranteed, given the unique human-machine interaction dynamics. We pilot-tested chatbot-based behaviour change through information provision and embedded animations. We evaluated whether the chatbot could increase understanding and intentions to adopt protective behaviours during the pandemic. Fifty-nine culturally and linguistically diverse participants received a compassion intervention, an exponential growth intervention, or no intervention. We measured participants\' COVID-19 testing intentions and measured their staying-home attitudes before and after their chatbot interaction. We found reduced uncertainty about protective behaviours. The exponential growth intervention increased participants\' testing intentions. This study provides preliminary evidence that chatbots can spark behaviour change, with applications in diverse and underrepresented groups.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    近年来,科学界已经被有趣的自主感觉经络反应(ASMR)迷住了,一种独特的现象,其特征是发源于头皮并沿脊柱向下传播的刺痛感。虽然轶事证据表明ASMR的治疗潜力,该领域见证了科学兴趣的激增,特别是通过使用神经成像技术,包括功能磁共振成像(fMRI)以及脑电图(EEG)和生理措施,如眼睛跟踪(瞳孔直径),心率(HR),心跳诱发电位(HEP),血压(BP),脉搏率(PR),手指光电体积描记术(PPG),和皮肤电导(SC)。本文旨在全面概述技术对科学阐明ASMR机制的贡献。
    进行了细致的文献综述,以确定使用EEG和生理测量检查ASMR的研究。全面搜索是在PUBMED等数据库中进行的,Scopus,IEEE,使用一系列相关关键字,如“ASMR”,“自主感觉经络反应”,\'脑电图\',\'fMRI\',\'脑电图\',\'生理测量\',\'心率\',\'皮肤电导\',和“眼动追踪”。这一严格的过程产生了大量的63篇PUBMED和166篇与SCOPUS相关的文章,确保在这篇综述中纳入广泛的高质量研究。
    这篇综述揭示了一系列利用脑电图和生理测量来探索ASMR效应的研究。脑电图研究揭示了与ASMR经历相关的不同大脑活动模式,特别是在涉及情绪加工和感觉统合的区域。在生理测量中,在ASMR触发的刺激期间,HR的降低以及SC和瞳孔直径的增加表明放松和注意力增加。
    这篇综述的发现强调了脑电图和生理措施在揭示ASMR的心理和生理影响方面的重要性。ASMR经验与独特的神经特征有关,而生理措施为ASMR刺激引起的自主反应提供了有价值的见解。这篇综述不仅强调了ASMR研究的跨学科性质,而且强调需要进一步研究以阐明ASMR的潜在机制并探索其潜在的治疗应用。从而为开发新的治疗干预措施铺平了道路。
    UNASSIGNED: In recent years, the scientific community has been captivated by the intriguing Autonomous sensory meridian response (ASMR), a unique phenomenon characterized by tingling sensations originating from the scalp and propagating down the spine. While anecdotal evidence suggests the therapeutic potential of ASMR, the field has witnessed a surge of scientific interest, particularly through the use of neuroimaging techniques including functional magnetic resonance imaging (fMRI) as well as electroencephalography (EEG) and physiological measures such as eye tracking (Pupil Diameter), heart rate (HR), heartbeat-evoked potential (HEP), blood pressure (BP), pulse rates (PR), finger photoplethysmography (PPG), and skin conductance (SC). This article is intended to provide a comprehensive overview of technology\'s contributions to the scientific elucidation of ASMR mechanisms.
    UNASSIGNED: A meticulous literature review was undertaken to identify studies that have examined ASMR using EEG and physiological measurements. The comprehensive search was conducted across databases such as PUBMED, SCOPUS, and IEEE, using a range of relevant keywords such as \'ASMR\', \'Autonomous sensory meridian response\', \'EEG\', \'fMRI\', \'electroencephalography\', \'physiological measures\', \'heart rate\', \'skin conductance\', and \'eye tracking\'. This rigorous process yielded a substantial number of 63 PUBMED and 166 SCOPUS-related articles, ensuring the inclusion of a wide range of high-quality research in this review.
    UNASSIGNED: The review uncovered a body of research utilizing EEG and physiological measures to explore ASMR\'s effects. EEG studies have revealed distinct patterns of brain activity associated with ASMR experiences, particularly in regions implicated in emotional processing and sensory integration. In physiological measurements, a decrease in HR and an increase in SC and pupil diameter indicate relaxation and increased attention during ASMR-triggered stimuli.
    UNASSIGNED: The findings of this review underscore the significance of EEG and physiological measures in unraveling the psychological and physiological effects of ASMR. ASMR experiences have been associated with unique neural signatures, while physiological measures provide valuable insights into the autonomic responses elicited by ASMR stimuli. This review not only highlights the interdisciplinary nature of ASMR research but also emphasizes the need for further investigation to elucidate the mechanisms underlying ASMR and explore its potential therapeutic applications, thereby paving the way for the development of novel therapeutic interventions.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    这项研究比较了在心理语言学实验中,讲英语的成年人和来自美国的儿童在与真人和聪明的演讲者(亚马逊Alexa)交谈时如何适应他们的演讲。总的来说,参与者在与设备交谈时产生更费力的演讲(持续时间更长,音调更高)。这些差异也因年龄而异:儿童在以设备为导向的语音中产生了更高的音调,暗示了更强烈的期望被系统误解。为了支持这一点,我们看到,在设备阶段性识别错误之后,孩子们加大了音高。此外,成人和儿童在回答“Alexa是否看起来像真人”时表现出相同程度的差异,进一步表明,儿童对系统能力的概念化塑造了他们的注册调整,而不是增加拟人化反应。这项工作谈到了语音产生背后机制的模型,和人机交互框架,为口语与技术互动的常规理论提供支持。
    This study compares how English-speaking adults and children from the United States adapt their speech when talking to a real person and a smart speaker (Amazon Alexa) in a psycholinguistic experiment. Overall, participants produced more effortful speech when talking to a device (longer duration and higher pitch). These differences also varied by age: children produced even higher pitch in device-directed speech, suggesting a stronger expectation to be misunderstood by the system. In support of this, we see that after a staged recognition error by the device, children increased pitch even more. Furthermore, both adults and children displayed the same degree of variation in their responses for whether \"Alexa seems like a real person or not\", further indicating that children\'s conceptualization of the system\'s competence shaped their register adjustments, rather than an increased anthropomorphism response. This work speaks to models on the mechanisms underlying speech production, and human-computer interaction frameworks, providing support for routinized theories of spoken interaction with technology.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    GPT-4的发布引起了各个领域的广泛关注,信号即将广泛采用和应用大型语言模型(LLM)。然而,以前的研究主要集中在ChatGPT的技术原理及其社会影响上,忽视了它对人机交互和用户心理的影响。本文探讨了ChatGPT对人机交互的多方面影响,心理学,和社会通过文献综述。作者调查了ChatGPT的技术基础,包括其Transformer架构和RLHF(来自人类反馈的强化学习)过程,使它能够产生类似人类的反应。在人机交互方面,作者研究了GPT模型给会话界面带来的重大改进。分析延伸到心理影响,权衡ChatGPT模仿人类同理心和支持学习的潜力,以减少人际关系的风险。在商业和社会领域,本文讨论了ChatGPT在客户服务和社会服务中的应用,强调效率的提高和隐私问题等挑战。最后,作者对ChatGPT的未来发展方向及其对社会关系的影响提供了预测和建议。
    The release of GPT-4 has garnered widespread attention across various fields, signaling the impending widespread adoption and application of Large Language Models (LLMs). However, previous research has predominantly focused on the technical principles of ChatGPT and its social impact, overlooking its effects on human-computer interaction and user psychology. This paper explores the multifaceted impacts of ChatGPT on human-computer interaction, psychology, and society through a literature review. The author investigates ChatGPT\'s technical foundation, including its Transformer architecture and RLHF (Reinforcement Learning from Human Feedback) process, enabling it to generate human-like responses. In terms of human-computer interaction, the author studies the significant improvements GPT models bring to conversational interfaces. The analysis extends to psychological impacts, weighing the potential of ChatGPT to mimic human empathy and support learning against the risks of reduced interpersonal connections. In the commercial and social domains, the paper discusses the applications of ChatGPT in customer service and social services, highlighting the improvements in efficiency and challenges such as privacy issues. Finally, the author offers predictions and recommendations for ChatGPT\'s future development directions and its impact on social relationships.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    这篇文章,从结构特征的角度来看,重点介绍车载用户界面图标,探讨不同图标结构特征对视觉搜索效率的影响。最初,我们根据结构特征将图标分为四组:个体结构图标(ISI),封闭结构图标(ESI),水平结构图标(HSI)和垂直结构图标(VSI)。随后,我们以结构为唯一变量进行了视觉搜索实验,记录参与者的行为和眼动追踪数据。最后,数据分析采用方差分析和逻辑回归分析.结果表明,图标结构特征的差异显著影响视觉搜索效率,显示显著的组间差异。恒生指数表现出最高的视觉搜索效率,而ESI显示效率最低。ISI的响应时间较短,但匹配精度最低。VSI的性能仅优于ESI。这些发现对优化图标设计和提高视觉搜索效率具有重要意义。
    图标的视觉搜索效率对于人机交互至关重要。我们研究了图标的结构特征如何影响视觉搜索效率。水平图标是最有效的,封闭的图标最少。单个图标是快速的,但不太准确。垂直图标优于封闭的图标。在设计中应考虑结构特征。
    This article, from the perspective of structural features, focuses on in-car user interface icons and explores the impact of different icon structural features on visual search efficiency. Initially, we categorised the icons into four groups based on structural features: individual structure icons (ISI), enclosed structure icons (ESI), horizontal structure icons (HSI) and vertical structure icons (VSI). Subsequently, we conducted a visual search experiment with structure as the sole variable, recording participants\' behaviours and eye-tracking data. Finally, data analysis was conducted using methods including analysis of variance and logistic regression. The results indicate that differences in icon structural features significantly affect visual search efficiency, showcasing significant intergroup differences. HSI exhibit the highest visual search efficiency, while ESI show the lowest efficiency. ISI have shorter response times but the lowest matching accuracy. VSI only perform better than ESI. These findings hold significant implications for optimising icon design and enhancing visual search efficiency.
    Visual search efficiency of icons is crucial for human-computer interaction. We investigated how the structural features of icons influence visual search efficiency. Horizontal icons are most effective, enclosed icons the least. Individual icons are quick but less accurate. Vertical icons outperform enclosed ones. Structural features should be considered in design.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    人类活动识别(HAR)在医疗保健和老年人保健(远程康复,远程监控),安全,人体工程学,娱乐(健身,体育推广,人机交互,视频游戏),和智能环境。本文解决了在运动训练中进行的12种练习的实时识别和重复计数问题。我们的方法基于深层神经网络模型,该模型由放置在胸部的9轴运动传感器(IMU)的信号提供。该模型可以在移动平台上运行(iOS,Android)。我们讨论了系统的设计要求及其对数据收集协议的影响。我们提出了基于预训练对比学习的编码器的体系结构。与端到端训练相比,所提出的方法在准确性方面显著提高了开发模型的质量(F1分数,MAPE)和背景活动期间的鲁棒性(假阳性率)。我们将AIDLAB-HAR数据集公开提供,以鼓励进一步的研究。
    Human Activity Recognition (HAR) plays an important role in the automation of various tasks related to activity tracking in such areas as healthcare and eldercare (telerehabilitation, telemonitoring), security, ergonomics, entertainment (fitness, sports promotion, human-computer interaction, video games), and intelligent environments. This paper tackles the problem of real-time recognition and repetition counting of 12 types of exercises performed during athletic workouts. Our approach is based on the deep neural network model fed by the signal from a 9-axis motion sensor (IMU) placed on the chest. The model can be run on mobile platforms (iOS, Android). We discuss design requirements for the system and their impact on data collection protocols. We present architecture based on an encoder pretrained with contrastive learning. Compared to end-to-end training, the presented approach significantly improves the developed model\'s quality in terms of accuracy (F1 score, MAPE) and robustness (false-positive rate) during background activity. We make the AIDLAB-HAR dataset publicly available to encourage further research.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    要了解人类推理与人工智能(AI)模型之间的一致性,这项实证研究将人类文本分类性能和可解释性与传统机器学习(ML)模型和大型语言模型(LLM)进行了比较。必须将包含204个伤害叙述的特定领域的嘈杂文本数据集分为6个伤害原因代码。根据伤害原因代码的独特性质,叙述的复杂性和分类难易程度各不相同。用户研究涉及51名参与者,他们在执行文本分类任务时记录了眼动跟踪数据。虽然ML模型是根据12万个预先标记的伤害叙述进行训练的,LLM和人类没有接受任何专业培训。根据用于进行分类决策的顶级单词,比较了不同方法的可解释性。这些单词是通过人类的眼动追踪来识别的,ML模型的可解释AI方法LIME,并提示LLM。观察到ML模型的分类性能相对优于零射LLM和非专业人类,总的来说,特别是对于高度复杂和难以分类的叙述。与后来的预测词相比,ML和LLM用于分类的前3个预测词在更大程度上与人类一致。
    To understand the alignment between reasonings of humans and artificial intelligence (AI) models, this empirical study compared the human text classification performance and explainability with a traditional machine learning (ML) model and large language model (LLM). A domain-specific noisy textual dataset of 204 injury narratives had to be classified into 6 cause-of-injury codes. The narratives varied in terms of complexity and ease of categorization based on the distinctive nature of cause-of-injury code. The user study involved 51 participants whose eye-tracking data was recorded while they performed the text classification task. While the ML model was trained on 120,000 pre-labelled injury narratives, LLM and humans did not receive any specialized training. The explainability of different approaches was compared based on the top words they used for making classification decision. These words were identified using eye-tracking for humans, explainable AI approach LIME for ML model, and prompts for LLM. The classification performance of ML model was observed to be relatively better than zero-shot LLM and non-expert humans, overall, and particularly for narratives with high complexity and difficult categorization. The top-3 predictive words used by ML and LLM for classification agreed with humans to a greater extent as compared to later predictive words.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

公众号