multimodal

多式联运
  • 文章类型: Journal Article
    BRCA1/2基因的状态在多种癌症类型的治疗决策过程中起着至关重要的作用。然而,由于高成本和有限的资源,目前,患者对BRCA1/2基因检测的需求尚未得到满足.值得注意的是,并非所有具有BRCA1/2突变的患者使用聚(ADP-核糖)聚合酶抑制剂(PARPi)均能获得良好的结局,表明风险分层的必要性。在这项研究中,我们旨在开发并验证预测BRCA1/2基因状态和PARPi治疗预后的多模式模型.
    我们纳入了1417例卵巢患者的1695张幻灯片,乳房,前列腺,和胰腺癌在三个独立的队列。利用自我注意机制,我们构建了一个多实例注意模型(MIAM),从苏木精和伊红(H&E)病理图像中检测BRCA1/2基因状态。我们进一步结合了MIAM模型的组织特征,细胞特征,和临床因素(MIAM-C模型)来预测BRCA1/2突变和PARPi治疗的无进展生存期(PFS)。使用曲线下面积(AUC)和Kaplan-Meier分析评价模型性能。分析了有助于MIAM-C的形态特征的可解释性。
    在四种癌症类型中,MIAM-C在识别BRCA1/2基因型方面优于基于深度学习的MIAM。可解释性分析显示,高度关注区域包括高级别肿瘤和淋巴细胞浸润,与BRCA1/2突变相关。值得注意的是,高淋巴细胞比率出现BRCA1/2突变的特征.此外,MIAM-C预测PARPi治疗反应(log-rankp<0.05),并作为BRCA1/2突变卵巢癌患者的独立预后因素(p<0.05,风险比:0.4,95%置信区间:0.16-0.99)。
    MIAM-C模型准确检测了BRCA1/2基因状态,并有效地对具有BRCA1/2突变的患者进行了分层预后。
    UNASSIGNED: The status of BRCA1/2 genes plays a crucial role in the treatment decision-making process for multiple cancer types. However, due to high costs and limited resources, a demand for BRCA1/2 genetic testing among patients is currently unmet. Notably, not all patients with BRCA1/2 mutations achieve favorable outcomes with poly (ADP-ribose) polymerase inhibitors (PARPi), indicating the necessity for risk stratification. In this study, we aimed to develop and validate a multimodal model for predicting BRCA1/2 gene status and prognosis with PARPi treatment.
    UNASSIGNED: We included 1695 slides from 1417 patients with ovarian, breast, prostate, and pancreatic cancers across three independent cohorts. Using a self-attention mechanism, we constructed a multi-instance attention model (MIAM) to detect BRCA1/2 gene status from hematoxylin and eosin (H&E) pathological images. We further combined tissue features from the MIAM model, cell features, and clinical factors (the MIAM-C model) to predict BRCA1/2 mutations and progression-free survival (PFS) with PARPi therapy. Model performance was evaluated using area under the curve (AUC) and Kaplan-Meier analysis. Morphological features contributing to MIAM-C were analyzed for interpretability.
    UNASSIGNED: Across the four cancer types, MIAM-C outperformed the deep learning-based MIAM in identifying the BRCA1/2 genotype. Interpretability analysis revealed that high-attention regions included high-grade tumors and lymphocytic infiltration, which correlated with BRCA1/2 mutations. Notably, high lymphocyte ratios appeared characteristic of BRCA1/2 mutations. Furthermore, MIAM-C predicted PARPi therapy response (log-rank p < 0.05) and served as an independent prognostic factor for patients with BRCA1/2-mutant ovarian cancer (p < 0.05, hazard ratio:0.4, 95% confidence interval: 0.16-0.99).
    UNASSIGNED: The MIAM-C model accurately detected BRCA1/2 gene status and effectively stratified prognosis for patients with BRCA1/2 mutations.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    先前基于静息态功能磁共振成像(rs-fMRI)和基于体素的形态计量学(VBM)的研究表明,早发性精神分裂症(EOS)患者的大脑结构和静息态功能大脑活动明显异常,与健康对照(HCs)相比,这些改变与EOS的发病机制密切相关。然而,以前的研究受到小样本量和结果高度异质性的限制。因此,本研究旨在有效整合以往的研究,以确定EOS患者常见和特定的脑功能和结构异常。
    PubMed,WebofScience,Embase,中国国家知识基础设施(CNKI),系统搜索了WanFang数据库,以确定有关EOS患者静息状态区域功能脑活动和灰质体积(GMV)异常的出版物。然后,我们利用基于种子的d映射与受试者图像排列(SDM-PSI)软件进行VBM和rs-fMRI研究的全脑体素荟萃分析,分别,并在此基础上进行多模态重叠,全面识别EOS患者的脑结构和功能异常。
    本荟萃分析共纳入了27项原始研究(28个数据集),包括与静息状态功能性脑活动相关的12项研究(13个数据集)(496名EOS患者,395项HCs)和15项与GMV相关的研究(15项数据集)(458例EOS患者,531HC)。总的来说,在功能荟萃分析中,EOS患者在左额中回(延伸至左额下回的三角形部分)和右尾状核中显示出静息状态的功能性脑活动显着增加。另一方面,在结构荟萃分析中,EOS患者在右颞上回(延伸到右罗兰骨)显示GMV显着降低,右颞中回,和颞极(颞上回)。
    这项荟萃分析显示,EOS中的某些区域表现出明显的结构或功能异常,比如时间回转,前额叶皮质,和纹状体。这些发现可能有助于加深我们对EOS潜在病理生理机制的理解,并为EOS的诊断或治疗提供潜在的生物标志物。
    UNASSIGNED: Previous studies based on resting-state functional magnetic resonance imaging(rs-fMRI) and voxel-based morphometry (VBM) have demonstrated significant abnormalities in brain structure and resting-state functional brain activity in patients with early-onset schizophrenia (EOS), compared with healthy controls (HCs), and these alterations were closely related to the pathogenesis of EOS. However, previous studies suffer from the limitations of small sample sizes and high heterogeneity of results. Therefore, the present study aimed to effectively integrate previous studies to identify common and specific brain functional and structural abnormalities in patients with EOS.
    UNASSIGNED: The PubMed, Web of Science, Embase, Chinese National Knowledge Infrastructure (CNKI), and WanFang databases were systematically searched to identify publications on abnormalities in resting-state regional functional brain activity and gray matter volume (GMV) in patients with EOS. Then, we utilized the Seed-based d Mapping with Permutation of Subject Images (SDM-PSI) software to conduct a whole-brain voxel meta-analysis of VBM and rs-fMRI studies, respectively, and followed by multimodal overlapping on this basis to comprehensively identify brain structural and functional abnormalities in patients with EOS.
    UNASSIGNED: A total of 27 original studies (28 datasets) were included in the present meta-analysis, including 12 studies (13 datasets) related to resting-state functional brain activity (496 EOS patients, 395 HCs) and 15 studies (15 datasets) related to GMV (458 EOS patients, 531 HCs). Overall, in the functional meta-analysis, patients with EOS showed significantly increased resting-state functional brain activity in the left middle frontal gyrus (extending to the triangular part of the left inferior frontal gyrus) and the right caudate nucleus. On the other hand, in the structural meta-analysis, patients with EOS showed significantly decreased GMV in the right superior temporal gyrus (extending to the right rolandic operculum), the right middle temporal gyrus, and the temporal pole (superior temporal gyrus).
    UNASSIGNED: This meta-analysis revealed that some regions in the EOS exhibited significant structural or functional abnormalities, such as the temporal gyri, prefrontal cortex, and striatum. These findings may help deepen our understanding of the underlying pathophysiological mechanisms of EOS and provide potential biomarkers for the diagnosis or treatment of EOS.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    机器学习(ML)在医疗人工智能(AI)系统中的应用已经从传统的统计方法转变为越来越多的深度学习模型应用。这项调查浏览了多模态ML的当前格局,重点关注其对医学图像分析和临床决策支持系统的深远影响。强调在解决多式联运代表方面的挑战和创新,聚变,翻译,对齐,和共同学习,本文探讨了多模态模型在临床预测中的转化潜力。它还强调需要对这种模式进行原则性评估和实际实施,关注决策支持系统与医疗保健提供者和人员之间的动态。尽管取得了进步,许多生物医学领域的数据偏见和“大数据”的稀缺等挑战依然存在。最后,我们讨论了原则性创新和协作努力,以进一步实现将多模态ML模型无缝集成到生物医学实践中的使命。
    Machine learning (ML) applications in medical artificial intelligence (AI) systems have shifted from traditional and statistical methods to increasing application of deep learning models. This survey navigates the current landscape of multimodal ML, focusing on its profound impact on medical image analysis and clinical decision support systems. Emphasizing challenges and innovations in addressing multimodal representation, fusion, translation, alignment, and co-learning, the paper explores the transformative potential of multimodal models for clinical predictions. It also highlights the need for principled assessments and practical implementation of such models, bringing attention to the dynamics between decision support systems and healthcare providers and personnel. Despite advancements, challenges such as data biases and the scarcity of \"big data\" in many biomedical domains persist. We conclude with a discussion on principled innovation and collaborative efforts to further the mission of seamless integration of multimodal ML models into biomedical practice.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    油茶是一种具有很高经济价值的作物,然而,它特别容易受到各种疾病和害虫的影响,这大大降低了它的产量和质量。因此,病害山茶叶的精确分割和分类对于有效管理病虫害至关重要。深度学习在植物病虫害分割方面表现出显著优势,特别是在复杂的图像处理和自动特征提取中。然而,在采用单模态模型分割油茶病害时,出现了三个关键挑战:(A)病变可能与复杂背景的颜色非常相似;(B)病叶的小部分重叠;(C)单叶上存在多种疾病。这些因素极大地阻碍了分割准确性。一种新颖的多模态模型,CNN-变压器双U形网络(CTDUNet),基于CNN-Transformer架构,已经被提出来集成图像和文本信息。该模型首先利用文本数据来解决单模态图像特征的缺点,增强其区分病变与环境特征的能力,即使是在彼此非常相似的条件下。此外,我们引入坐标空间注意力(CSA),它专注于目标之间的位置关系,从而改善了重叠叶边的分割。此外,交叉注意力(CA)用于有效地对齐图像和文本特征,保存本地信息,增强对各种疾病的感知和区分。CTDUNet模型在自制的多模态数据集上进行了评估,并与几个模型进行了比较,包括DeeplabV3+,UNet,PSPNet,Segformer,HrNet,和语言满足视觉转换(LVIT)。实验结果表明,CTDUNet实现了86.14%的平均交集(mIoU),分别超过多模态模型和最佳单模态模型3.91%和5.84%,分别。此外,CTDUNet在油茶病虫害的多分类中表现出高度平衡。这些结果表明融合图像和文本多模态信息在山茶病分割中的成功应用,实现了出色的性能。
    Camellia oleifera is a crop of high economic value, yet it is particularly susceptible to various diseases and pests that significantly reduce its yield and quality. Consequently, the precise segmentation and classification of diseased Camellia leaves are vital for managing pests and diseases effectively. Deep learning exhibits significant advantages in the segmentation of plant diseases and pests, particularly in complex image processing and automated feature extraction. However, when employing single-modal models to segment Camellia oleifera diseases, three critical challenges arise: (A) lesions may closely resemble the colors of the complex background; (B) small sections of diseased leaves overlap; (C) the presence of multiple diseases on a single leaf. These factors considerably hinder segmentation accuracy. A novel multimodal model, CNN-Transformer Dual U-shaped Network (CTDUNet), based on a CNN-Transformer architecture, has been proposed to integrate image and text information. This model first utilizes text data to address the shortcomings of single-modal image features, enhancing its ability to distinguish lesions from environmental characteristics, even under conditions where they closely resemble one another. Additionally, we introduce Coordinate Space Attention (CSA), which focuses on the positional relationships between targets, thereby improving the segmentation of overlapping leaf edges. Furthermore, cross-attention (CA) is employed to align image and text features effectively, preserving local information and enhancing the perception and differentiation of various diseases. The CTDUNet model was evaluated on a self-made multimodal dataset compared against several models, including DeeplabV3+, UNet, PSPNet, Segformer, HrNet, and Language meets Vision Transformer (LViT). The experimental results demonstrate that CTDUNet achieved an mean Intersection over Union (mIoU) of 86.14%, surpassing both multimodal models and the best single-modal model by 3.91% and 5.84%, respectively. Additionally, CTDUNet exhibits high balance in the multi-class segmentation of Camellia oleifera diseases and pests. These results indicate the successful application of fused image and text multimodal information in the segmentation of Camellia disease, achieving outstanding performance.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    脑机接口(BCI)是神经科学中最强大的工具之一,通常包括一个记录系统,处理器系统,和一个刺激系统。光遗传学具有双向调节的优势,高时空分辨率,和细胞特异性调节,扩展了BCI的应用场景。近年来,随着材料和软件的发展,光遗传学BCI已广泛用于实验室。这些系统被设计得更加集成,轻量级,生物相容性和高效的电源,无线传输和芯片级嵌入式BCI也是如此。软件也在不断改进,具有更好的实时性能和准确性以及更低的功耗。另一方面,作为一项跨越分子生物学等多学科领域的尖端技术,神经科学,材料工程,和信息处理,光遗传学BCI在神经解码中具有巨大的应用潜力,增强大脑功能,治疗神经疾病。这里,本文综述了光遗传学BCIs的发展和应用。在未来,结合其他功能成像技术,如近红外光谱(fNIRS)和功能磁共振成像(fMRI),光遗传学BCI可以调节特定电路的功能,促进神经康复,协助感知,建立脑-脑接口,并应用于更广泛的应用场景。
    The brain-computer interface (BCI) is one of the most powerful tools in neuroscience and generally includes a recording system, a processor system, and a stimulation system. Optogenetics has the advantages of bidirectional regulation, high spatiotemporal resolution, and cell-specific regulation, which expands the application scenarios of BCIs. In recent years, optogenetic BCIs have become widely used in the lab with the development of materials and software. The systems were designed to be more integrated, lightweight, biocompatible, and power efficient, as were the wireless transmission and chip-level embedded BCIs. The software is also constantly improving, with better real-time performance and accuracy and lower power consumption. On the other hand, as a cutting-edge technology spanning multidisciplinary fields including molecular biology, neuroscience, material engineering, and information processing, optogenetic BCIs have great application potential in neural decoding, enhancing brain function, and treating neural diseases. Here, we review the development and application of optogenetic BCIs. In the future, combined with other functional imaging techniques such as near-infrared spectroscopy (fNIRS) and functional magnetic resonance imaging (fMRI), optogenetic BCIs can modulate the function of specific circuits, facilitate neurological rehabilitation, assist perception, establish a brain-to-brain interface, and be applied in wider application scenarios.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    近年来,随着X射线等各种成像模态的集成,医学图像配准已变得至关重要。超声,MRI,还有CT扫描,能够全面分析和诊断生物结构。本文对医学图像配准技术进行了全面的综述,深入关注2D-2D图像配准方法。虽然简要地谈到了3D注册,主要重点仍然是2D技术及其应用。本综述涵盖了各种模式的注册技术,包括单峰,多模态,患者间,和患者内部。本文探讨了医学图像配准中遇到的挑战,包括几何失真,图像属性的差异,异常值,和优化收敛,并讨论了它们对配准精度和可靠性的影响。强调了应对这些挑战的战略,强调需要不断创新和改进技术,以提高医学图像配准系统的准确性和可靠性。最后,本文强调了准确的医学图像配准在改善诊断中的重要性。
    Medical image registration has become pivotal in recent years with the integration of various imaging modalities like X-ray, ultrasound, MRI, and CT scans, enabling comprehensive analysis and diagnosis of biological structures. This paper provides a comprehensive review of registration techniques for medical images, with an in-depth focus on 2D-2D image registration methods. While 3D registration is briefly touched upon, the primary emphasis remains on 2D techniques and their applications. This review covers registration techniques for diverse modalities, including unimodal, multimodal, interpatient, and intra-patient. The paper explores the challenges encountered in medical image registration, including geometric distortion, differences in image properties, outliers, and optimization convergence, and discusses their impact on registration accuracy and reliability. Strategies for addressing these challenges are highlighted, emphasizing the need for continual innovation and refinement of techniques to enhance the accuracy and reliability of medical image registration systems. The paper concludes by emphasizing the importance of accurate medical image registration in improving diagnosis.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    本研究旨在探索混合深度学习和影像组学方法的功效,补充了患者元数据,在基于非侵入性皮肤镜成像的皮肤病变诊断中。我们分析了来自国际皮肤成像合作组织(ISIC)数据集的皮肤镜图像,跨越2016-2020年,涵盖各种皮肤病变。我们的方法将深度学习与全面的影像组学分析相结合,利用大量的定量图像特征来精确量化皮肤病变模式。数据集包括三个案例,四,和八种不同的皮肤损伤类型。我们的方法以ISIC2020挑战和先前研究中使用二元决策框架的七种分类方法为基准。提出的混合模型在区分良性和恶性病变方面表现出优异的性能,实现99%的受试者工作特征曲线下面积(AUROC)评分,95%,96%,多类解码AUROC为98.5%,94.9%,和96.4%,敏感度为97.6%,93.9%,96.0%和98.4%的特异性,96.7%,在2018年ISIC内部挑战中占96.9%,以及在济南和龙华的外部数据集中,分别。我们的研究结果表明,影像组学和深度学习的整合,利用皮肤镜图像,有效捕获皮肤病变的异质性和模式表达。
    This study aims to explore the efficacy of a hybrid deep learning and radiomics approach, supplemented with patient metadata, in the noninvasive dermoscopic imaging-based diagnosis of skin lesions. We analyzed dermoscopic images from the International Skin Imaging Collaboration (ISIC) dataset, spanning 2016-2020, encompassing a variety of skin lesions. Our approach integrates deep learning with a comprehensive radiomics analysis, utilizing a vast array of quantitative image features to precisely quantify skin lesion patterns. The dataset includes cases of three, four, and eight different skin lesion types. Our methodology was benchmarked against seven classification methods from the ISIC 2020 challenge and prior research using a binary decision framework. The proposed hybrid model demonstrated superior performance in distinguishing benign from malignant lesions, achieving area under the receiver operating characteristic curve (AUROC) scores of 99%, 95%, and 96%, and multiclass decoding AUROCs of 98.5%, 94.9%, and 96.4%, with sensitivities of 97.6%, 93.9%, and 96.0% and specificities of 98.4%, 96.7%, and 96.9% in the internal ISIC 2018 challenge, as well as in the external Jinan and Longhua datasets, respectively. Our findings suggest that the integration of radiomics and deep learning, utilizing dermoscopic images, effectively captures the heterogeneity and pattern expression of skin lesions.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    在产房,胎儿的健康状况是通过实验室检查来评估的,像心脏描记术这样的生物信号,和成像技术,如胎儿超声心动图。我们开发了一种多模态机器学习模型,该模型集成了医疗记录,生物信号,和影像学数据来预测胎儿酸中毒,使用三级医院产房的数据集(n=2266)。为了实现这一点,特征是从非结构化数据源中提取的,包括生物信号和成像,然后与医疗记录中的结构化数据合并。连接的向量形成用于训练分类器以预测分娩后胎儿酸中毒的基础。我们的模型在测试数据集上实现了0.752的接收器工作特征曲线下面积(AUROC),证明了多模式模型在预测各种胎儿结局方面的潜力。
    In the delivery room, fetal well-being is evaluated through laboratory tests, biosignals like cardiotocography, and imaging techniques such as fetal echocardiography. We have developed a multimodal machine learning model that integrates medical records, biosignals, and imaging data to predict fetal acidosis, using a dataset from a tertiary hospital\'s delivery room (n=2,266). To achieve this, features were extracted from unstructured data sources, including biosignals and imaging, and then merged with structured data from medical records. The concatenated vectors formed the basis for training a classifier to predict post-delivery fetal acidosis. Our model achieved an Area Under the Receiver Operating Characteristic curve (AUROC) of 0.752 on the test dataset, demonstrating the potential of multimodal models in predicting various fetal outcomes.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    为了放射科医生之间更好的合作,应评估口译工作量,考虑到解释每种情况的难度差异。然而,客观评估难度是具有挑战性的。这项研究提出了一种结构和文本数据的多模态分类器,以在不使用图像的情况下根据订单信息和患者数据预测难度。分类器显示特异性为0.9和准确度为0.7的性能。
    For better collaboration among radiologists, the interpretation workload should be evaluated, considering the difference in difficulty for interpreting each case. However, objective evaluation of difficulty is challenging. This study proposes a multimodal classifier of structural and textual data to predict difficulty based on order information and patient data without using images. The classifier showed performance with a specificity of 0.9 and an accuracy of 0.7.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    单细胞多模态数据集测量了单个细胞的各种特征,使细胞和分子机制的深刻理解。然而,多模态数据生成仍然昂贵且具有挑战性,缺失的模式经常发生。最近,机器学习方法已经被开发用于数据插补,但通常需要完全匹配的多模态来学习可能缺乏模态特异性的常见潜在嵌入。为了解决这些问题,我们开发了一个开源的机器学习模型,用于多模态填充和嵌入的联合变分自编码器(JAMIE)。JAMIE获取单细胞多模态数据,这些数据可以具有跨模态的部分匹配样本。变分自动编码器学习每种模态的潜在嵌入。然后,来自跨模态匹配样本的嵌入在重建之前被聚合以识别联合跨模态潜在嵌入。要执行跨模态插补,一个模态的潜在嵌入可以与另一个模态的解码器一起使用。为了可解释性,Shapley值用于对跨模态插补和已知样本标签的输入特征进行优先级排序。我们将JAMIE应用于模拟数据和新兴的单细胞多模态数据,包括基因表达,染色质可及性,人类和小鼠大脑的电生理学。JAMIE在一般情况下显著优于现有的最先进的方法和优先考虑的多模态特征,在细胞分辨率方面提供潜在的新颖机械见解。
    Single-cell multimodal datasets have measured various characteristics of individual cells, enabling a deep understanding of cellular and molecular mechanisms. However, multimodal data generation remains costly and challenging, and missing modalities happen frequently. Recently, machine learning approaches have been developed for data imputation but typically require fully matched multimodalities to learn common latent embeddings that potentially lack modality specificity. To address these issues, we developed an open-source machine learning model, Joint Variational Autoencoders for multimodal Imputation and Embedding (JAMIE). JAMIE takes single-cell multimodal data that can have partially matched samples across modalities. Variational autoencoders learn the latent embeddings of each modality. Then, embeddings from matched samples across modalities are aggregated to identify joint cross-modal latent embeddings before reconstruction. To perform cross-modal imputation, the latent embeddings of one modality can be used with the decoder of the other modality. For interpretability, Shapley values are used to prioritize input features for cross-modal imputation and known sample labels. We applied JAMIE to both simulation data and emerging single-cell multimodal data including gene expression, chromatin accessibility, and electrophysiology in human and mouse brains. JAMIE significantly outperforms existing state-of-the-art methods in general and prioritized multimodal features for imputation, providing potentially novel mechanistic insights at cellular resolution.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

公众号