Explainability

可解释性
  • 文章类型: Journal Article
    中风是死亡和发病的重要原因,需要早期预测策略以最大限度地降低风险。传统的评估患者的方法,如急性生理学和慢性健康评估(APACHEII,IV)和简化急性生理学评分III(SAPSIII),具有有限的准确性和可解释性。本文提出了一种新颖的方法:一种可解释的方法,基于注意力的早期卒中死亡率预测变压器模型。该模型旨在解决以前预测模型的局限性,同时提供可解释性(提供明确的,模型的可理解解释)和保真度(给出模型从输入到输出的动态的真实解释)。此外,本研究使用Shapley值和基于注意力的评分来探索和比较保真度和可解释性评分,以提高模型的可解释性.研究目标包括设计一个可解释的基于注意力的变压器模型,与现有模型相比,评估其性能,并提供从模型得出的特征重要性。
    Stroke is a significant cause of mortality and morbidity, necessitating early predictive strategies to minimize risks. Traditional methods for evaluating patients, such as Acute Physiology and Chronic Health Evaluation (APACHE II, IV) and Simplified Acute Physiology Score III (SAPS III), have limited accuracy and interpretability. This paper proposes a novel approach: an interpretable, attention-based transformer model for early stroke mortality prediction. This model seeks to address the limitations of previous predictive models, providing both interpretability (providing clear, understandable explanations of the model) and fidelity (giving a truthful explanation of the model\'s dynamics from input to output). Furthermore, the study explores and compares fidelity and interpretability scores using Shapley values and attention-based scores to improve model explainability. The research objectives include designing an interpretable attention-based transformer model, evaluating its performance compared to existing models, and providing feature importance derived from the model.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    可解释性是提高人工智能在医学中可信度的关键。然而,医生对模型可解释性的期望与这些模型的实际行为之间存在显著差距。这种差距是由于缺乏以医生为中心的评估框架的共识。需要定量评估有效的可解释性应为从业者提供的实际利益。这里,我们假设优越的注意力映射,作为一种模型解释的机制,应该与医生关注的信息保持一致,潜在地降低预测不确定性并提高模型可靠性。我们使用多模式变压器使用临床数据和磁共振成像来预测直肠癌的淋巴结转移。我们探索了注意力地图有多好,通过最先进的技术可视化,可以与医生的理解达成一致。随后,我们比较了两种不同的不确定性估计方法:仅使用预测概率方差的独立估计,以及考虑预测概率方差和量化一致性的人在环估计。我们的发现表明,与独立方法相比,人在环方法没有显着优势。总之,本案例研究未证实该解释在增强模型可靠性方面的预期益处.肤浅的解释可能弊大于利,误导医生依赖不确定的预测,这表明,在模型可解释性的背景下,不应高估注意力机制的当前状态。
    Explainability is key to enhancing the trustworthiness of artificial intelligence in medicine. However, there exists a significant gap between physicians\' expectations for model explainability and the actual behavior of these models. This gap arises from the absence of a consensus on a physician-centered evaluation framework, which is needed to quantitatively assess the practical benefits that effective explainability should offer practitioners. Here, we hypothesize that superior attention maps, as a mechanism of model explanation, should align with the information that physicians focus on, potentially reducing prediction uncertainty and increasing model reliability. We employed a multimodal transformer to predict lymph node metastasis of rectal cancer using clinical data and magnetic resonance imaging. We explored how well attention maps, visualized through a state-of-the-art technique, can achieve agreement with physician understanding. Subsequently, we compared two distinct approaches for estimating uncertainty: a standalone estimation using only the variance of prediction probability, and a human-in-the-loop estimation that considers both the variance of prediction probability and the quantified agreement. Our findings revealed no significant advantage of the human-in-the-loop approach over the standalone one. In conclusion, this case study did not confirm the anticipated benefit of the explanation in enhancing model reliability. Superficial explanations could do more harm than good by misleading physicians into relying on uncertain predictions, suggesting that the current state of attention mechanisms should not be overestimated in the context of model explainability.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    背景:不良事件是指在医院对患者有潜在或实际伤害的事件。这些事件通常通过患者安全事件(PSE)报告进行记录。其中包括详细的叙述,提供有关事件的上下文信息。PSE报告的准确分类对于患者安全监测至关重要。然而,由于分类不一致和报告数量庞大,这一过程面临挑战。文本表示的最新进展,特别是从基于转换器的语言模型派生的上下文文本表示,为更精确的PSE报告分类提供了一个有前途的解决方案。集成机器学习(ML)分类器需要在人类专业知识和人工智能(AI)之间取得平衡。这种整合的核心是可解释性的概念,这对于建立信任和确保有效的人与人工智能协作至关重要。
    目的:本研究旨在研究使用上下文文本表示训练的ML分类器在自动分类PSE报告中的功效。此外,该研究提出了一个界面,该界面将ML分类器与可解释性技术集成在一起,以促进PSE报告分类的人与人工智能协作。
    方法:本研究使用了来自美国东南部一家大型学术医院产科的861份PSE报告的数据集。使用PSE报告的静态和上下文文本表示来训练各种ML分类器。使用多类分类度量和混淆矩阵评估训练的ML分类器。使用本地可解释模型不可知解释(LIME)技术来提供ML分类器预测的基本原理。为事件报告系统设计了将ML分类器与LIME技术集成的接口。
    结果:使用上下文表示的最佳分类器能够获得75.4%(95/126)的准确性,而使用静态文本表示训练的最佳分类器的准确性为66.7%(84/126)。已设计了PSE报告界面,以促进PSE报告分类中的人类与AI协作。在这个设计中,ML分类器推荐前2个最可能的事件类型,以及对预测的解释,使PSE记者和患者安全分析师选择最合适的一个。LIME技术表明,分类器偶尔依赖于任意单词进行分类,强调人类监督的必要性。
    结论:这项研究表明,使用上下文文本表示训练ML分类器可以显着提高PSE报告分类的准确性。本研究设计的界面为PSE报告分类中的人与人协作奠定了基础。从这项研究中获得的见解增强了PSE报告分类中的决策过程,使医院能够更有效地识别潜在的风险和危害,并使患者安全分析师能够及时采取行动,防止患者受到伤害。
    BACKGROUND: Adverse events refer to incidents with potential or actual harm to patients in hospitals. These events are typically documented through patient safety event (PSE) reports, which consist of detailed narratives providing contextual information on the occurrences. Accurate classification of PSE reports is crucial for patient safety monitoring. However, this process faces challenges due to inconsistencies in classifications and the sheer volume of reports. Recent advancements in text representation, particularly contextual text representation derived from transformer-based language models, offer a promising solution for more precise PSE report classification. Integrating the machine learning (ML) classifier necessitates a balance between human expertise and artificial intelligence (AI). Central to this integration is the concept of explainability, which is crucial for building trust and ensuring effective human-AI collaboration.
    OBJECTIVE: This study aims to investigate the efficacy of ML classifiers trained using contextual text representation in automatically classifying PSE reports. Furthermore, the study presents an interface that integrates the ML classifier with the explainability technique to facilitate human-AI collaboration for PSE report classification.
    METHODS: This study used a data set of 861 PSE reports from a large academic hospital\'s maternity units in the Southeastern United States. Various ML classifiers were trained with both static and contextual text representations of PSE reports. The trained ML classifiers were evaluated with multiclass classification metrics and the confusion matrix. The local interpretable model-agnostic explanations (LIME) technique was used to provide the rationale for the ML classifier\'s predictions. An interface that integrates the ML classifier with the LIME technique was designed for incident reporting systems.
    RESULTS: The top-performing classifier using contextual representation was able to obtain an accuracy of 75.4% (95/126) compared to an accuracy of 66.7% (84/126) by the top-performing classifier trained using static text representation. A PSE reporting interface has been designed to facilitate human-AI collaboration in PSE report classification. In this design, the ML classifier recommends the top 2 most probable event types, along with the explanations for the prediction, enabling PSE reporters and patient safety analysts to choose the most suitable one. The LIME technique showed that the classifier occasionally relies on arbitrary words for classification, emphasizing the necessity of human oversight.
    CONCLUSIONS: This study demonstrates that training ML classifiers with contextual text representations can significantly enhance the accuracy of PSE report classification. The interface designed in this study lays the foundation for human-AI collaboration in the classification of PSE reports. The insights gained from this research enhance the decision-making process in PSE report classification, enabling hospitals to more efficiently identify potential risks and hazards and enabling patient safety analysts to take timely actions to prevent patient harm.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    近年来,与公共卫生相关的公开数据的可用性显着增加。这些数据具有制定公共卫生政策的巨大潜力;然而,这需要有意义和深刻的分析。我们的目的是演示如何使用数据分析技术来解决数据减少的问题,使用在线可用的公共卫生数据进行预测和解释,以便为公共卫生政策提供良好的依据。
    从现有的在线英国国家公共卫生数据库中分析了观察性自杀预防数据。多重共线性分析和主成分分析用于减少相关数据,其次是预测和解释自杀的回归分析。
    多重共线性分析有效地将预测因子的指标集减少了30%,而主成分分析进一步将指标集减少了86%。用于预测的回归确定了自杀行为的四个重要指标预测因子(因故意自残而紧急住院,离开照顾的孩子,法定无家可归和自我报告的幸福感/低幸福感)和两个主要成分预测因子(相关度功能障碍,以及行为问题和精神疾病)。对解释的回归确定了社会因素(独自生活)对自杀行为的幸福感(低幸福感)的显着调节作用,从而支持现有的理论,并提供超越回归预测结果的洞察力。还确定了两个独立的预测因子,这些预测因子捕获了社会护理服务提供中的相关性需求。
    我们证明了回归技术在在线公共卫生数据分析中的有效性。预测和解释的回归分析都适用于公共卫生数据分析,以便更好地了解公共卫生结果。因此,必须明确分析的目的(预测准确性或理论发展),作为选择最合适模型的基础。我们将这些技术应用于自杀数据分析;然而,我们认为,本研究中提出的分析应应用于整个公共卫生领域的数据集,以提高卫生政策建议的质量.
    In recent years, the availability of publicly available data related to public health has significantly increased. These data have substantial potential to develop public health policy; however, this requires meaningful and insightful analysis. Our aim is to demonstrate how data analysis techniques can be used to address the issues of data reduction, prediction and explanation using online available public health data, in order to provide a sound basis for informing public health policy.
    Observational suicide prevention data were analysed from an existing online United Kingdom national public health database. Multi-collinearity analysis and principal-component analysis were used to reduce correlated data, followed by regression analyses for prediction and explanation of suicide.
    Multi-collinearity analysis was effective in reducing the indicator set of predictors by 30% and principal component analysis further reduced the set by 86%. Regression for prediction identified four significant indicator predictors of suicide behaviour (emergency hospital admissions for intentional self-harm, children leaving care, statutory homelessness and self-reported well-being/low happiness) and two main component predictors (relatedness dysfunction, and behavioural problems and mental illness). Regression for explanation identified significant moderation of a well-being predictor (low happiness) of suicide behaviour by a social factor (living alone), thereby supporting existing theory and providing insight beyond the results of regression for prediction. Two independent predictors capturing relatedness needs in social care service delivery were also identified.
    We demonstrate the effectiveness of regression techniques in the analysis of online public health data. Regression analysis for prediction and explanation can both be appropriate for public health data analysis for a better understanding of public health outcomes. It is therefore essential to clarify the aim of the analysis (prediction accuracy or theory development) as a basis for choosing the most appropriate model. We apply these techniques to the analysis of suicide data; however, we argue that the analysis presented in this study should be applied to datasets across public health in order to improve the quality of health policy recommendations.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Sci-hub)

       PDF(Pubmed)

公众号