Multiple instance learning

多实例学习
  • 文章类型: Journal Article
    通过计算机辅助诊断的自动宫颈癌筛查已显示出改善筛查可及性并减少相关成本和错误的巨大潜力。然而,由于患者特定的变化,整个幻灯片图像(WSI)的分类性能仍然次优。为了提高筛选的精度,病理学家不仅分析可疑异常细胞的特征,还可以将它们与正常细胞进行比较。受这种做法的激励,我们提出了一种新的宫颈细胞比较学习方法,该方法利用病理学家的知识来学习同一WSI中正常细胞和可疑异常细胞之间的差异。我们的方法采用两个预先训练的YOLOX模型来检测给定WSI中的可疑异常和正常细胞。然后,自监督模型提取检测到的细胞的特征。随后,定制的Transformer编码器将单元特征融合以获得WSI实例嵌入。最后,基于注意力的多实例学习实现分类。实验结果表明,我们提出的方法的AUC为0.9319。此外,该方法达到了专业病理学家水平的性能,表明其临床应用的潜力。
    Automated cervical cancer screening through computer-assisted diagnosis has shown considerable potential to improve screening accessibility and reduce associated costs and errors. However, classification performance on whole slide images (WSIs) remains suboptimal due to patient-specific variations. To improve the precision of the screening, pathologists not only analyze the characteristics of suspected abnormal cells, but also compare them with normal cells. Motivated by this practice, we propose a novel cervical cell comparative learning method that leverages pathologist knowledge to learn the differences between normal and suspected abnormal cells within the same WSI. Our method employs two pre-trained YOLOX models to detect suspected abnormal and normal cells in a given WSI. A self-supervised model then extracts features for the detected cells. Subsequently, a tailored Transformer encoder fuses the cell features to obtain WSI instance embeddings. Finally, attention-based multi-instance learning is applied to achieve classification. The experimental results show an AUC of 0.9319 for our proposed method. Moreover, the method achieved professional pathologist-level performance, indicating its potential for clinical applications.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    目的慢性血栓栓塞性肺动脉高压(CTEPH)的早期非特异性症状诊断具有挑战性。复杂的诊断过程,和小的病变大小。本研究旨在开发一种使用非对比计算机断层扫描(NCCT)扫描的CTEPH自动诊断方法,无需精确的病变注释即可实现自动诊断。
方法开发了一种具有多实例学习(CNMIL)框架的新型级联网络,以改善CTEPH的诊断。该方法使用结合两个Resnet-18CNN网络的级联网络架构来逐步区分正常情况和CTEPH情况。多实例学习(MIL)用于将每个3DCT病例视为图像切片的“袋子”,使用注意力评分来识别最重要的切片。注意模块帮助模型专注于每个切片内的诊断相关区域。数据集包括来自300名受试者的NCCT扫描,包括117名男性和183名女性,平均年龄为52.5±20.9岁,包括132例正常病例和168例肺部疾病,包括88例CTEPH。CNMIL框架使用灵敏度进行了评估,特异性,和曲线下面积(AUC)指标,并与常见的3D监督分类网络和现有的CTEPH自动诊断网络进行了比较。 主要结果CNMIL框架显示出高诊断性能,在区分CTEPH病例时,AUC为0.807,准确性为0.833,敏感性为0.795,特异性为0.849。消融研究表明,集成MIL和级联网络显着增强了性能,该模型在正常分类中达到0.993的AUC和完美的灵敏度(1.000)。与其他3D网络体系结构的比较证实,集成模型优于其他模型,达到0.8419的最高AUC。 意义CNMIL网络不需要额外的扫描或注释,完全依靠NCCT。这种方法可以提高CTEPH检测的及时性和准确性,导致更好的患者结果。
    ObjectiveThe diagnosis of chronic thromboembolic pulmonary hypertension (CTEPH) is challenging due to nonspecific early symptoms, complex diagnostic processes, and small lesion sizes. This study aims to develop an automatic diagnosis method for CTEPH using non-contrasted computed tomography (NCCT) scans, enabling automated diagnosis without precise lesion annotation. ApproachA novel Cascade Network with Multiple Instance Learning (CNMIL) framework was developed to improve the diagnosis of CTEPH. This method uses a cascade network architecture combining two Resnet-18 CNN networks to progressively distinguish between normal and CTEPH cases. Multiple Instance Learning (MIL) is employed to treat each 3D CT case as a \"bag\" of image slices, using attention scoring to identify the most important slices. An attention module helps the model focus on diagnostically relevant regions within each slice. The dataset comprised NCCT scans from 300 subjects, including 117 males and 183 females, with an average age of 52.5 ± 20.9 years, consisting of 132 normal cases and 168 cases of lung diseases, including 88 instances of CTEPH. The CNMIL framework was evaluated using sensitivity, specificity, and the area under the curve (AUC) metrics, and compared with common 3D supervised classification networks and existing CTEPH automatic diagnosis networks. Main ResultsThe CNMIL framework demonstrated high diagnostic performance, achieving an AUC of 0.807, accuracy of 0.833, sensitivity of 0.795, and specificity of 0.849 in distinguishing CTEPH cases. Ablation studies revealed that integrating MIL and the cascade network significantly enhanced performance, with the model achieving an AUC of 0.993 and perfect sensitivity (1.000) in normal classification. Comparisons with other 3D network architectures confirmed that the integrated model outperformed others, achieving the highest AUC of 0.8419. SignificanceThe CNMIL network requires no additional scans or annotations, relying solely on NCCT. This approach can improve timely and accurate CTEPH detection, resulting in better patient outcomes.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    T细胞受体(TCR)是人体免疫系统的关键,了解其细微差别可以显着增强我们预测癌症相关免疫反应的能力。然而,现有的方法经常忽略T细胞受体(TCR)的序列内和序列间相互作用,限制了基于序列的癌症相关免疫状态预测的发展。为了应对这一挑战,我们提议BertTCR,一个创新的深度学习框架,旨在使用TCR预测癌症相关的免疫状态。BertTCR将预先训练的蛋白质大语言模型与深度学习架构相结合,使其能够从TCR中提取更深层次的上下文信息。与三种最先进的基于序列的方法相比,BertTCR将甲状腺癌检测的外部验证集的AUC提高了21个百分点。此外,该模型在2000多个公开可用的TCR库上进行了训练,涵盖了17种癌症和健康样本,它已经在多个公共外部数据集上验证了其区分癌症患者和健康个体的能力。此外,BertTCR可以准确地对各种癌症类型和健康个体进行分类。总的来说,BertTCR是基于TCR的癌症相关免疫状态预测的先进方法,为广泛的免疫状态预测任务提供了有希望的潜力。
    The T cell receptor (TCR) repertoire is pivotal to the human immune system, and understanding its nuances can significantly enhance our ability to forecast cancer-related immune responses. However, existing methods often overlook the intra- and inter-sequence interactions of T cell receptors (TCRs), limiting the development of sequence-based cancer-related immune status predictions. To address this challenge, we propose BertTCR, an innovative deep learning framework designed to predict cancer-related immune status using TCRs. BertTCR combines a pre-trained protein large language model with deep learning architectures, enabling it to extract deeper contextual information from TCRs. Compared to three state-of-the-art sequence-based methods, BertTCR improves the AUC on an external validation set for thyroid cancer detection by 21 percentage points. Additionally, this model was trained on over 2000 publicly available TCR libraries covering 17 types of cancer and healthy samples, and it has been validated on multiple public external datasets for its ability to distinguish cancer patients from healthy individuals. Furthermore, BertTCR can accurately classify various cancer types and healthy individuals. Overall, BertTCR is the advancing method for cancer-related immune status forecasting based on TCRs, offering promising potential for a wide range of immune status prediction tasks.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    在计算病理学领域中,基于多实例学习(MIL)的方法已经被广泛采用来处理整个幻灯片图像(WSI)。由于幻灯片级监督稀疏,这些方法通常在肿瘤区域缺乏良好的定位,导致可解释性差。此外,它们缺乏对预测结果的稳健不确定性估计,导致可靠性差。为了解决上述两个限制,我们提出了一个可解释和证据的多实例学习(E2-MIL)框架,用于整个幻灯片图像分类。E2-MIL主要由三个模块组成:细节感知注意蒸馏模块(DAM),结构感知注意力细化模块(SRM),和不确定性感知实例分类器(UIC)。具体来说,DAM通过利用互补的子袋从本地网络中学习详细的注意力知识,帮助全球网络找到更多细节感知的正面实例。此外,还引入了屏蔽的自指导损失,以帮助弥合幻灯片级别标签和实例级别分类任务之间的差距。SRM生成结构感知注意力图,其通过有效地对聚类实例之间的空间关系建模来定位整个肿瘤区域结构。此外,UIC提供准确的实例级分类结果和稳健的预测不确定性估计,以提高基于主观逻辑理论的模型可靠性。在三个大型多中心子类型数据集上进行的大量实验证明了E2-MIL的幻灯片级和实例级性能优势。
    Multiple instance learning (MIL)-based methods have been widely adopted to process the whole slide image (WSI) in the field of computational pathology. Due to the sparse slide-level supervision, these methods usually lack good localization on the tumor regions, leading to poor interpretability. Moreover, they lack robust uncertainty estimation of prediction results, leading to poor reliability. To solve the above two limitations, we propose an explainable and evidential multiple instance learning (E2-MIL) framework for whole slide image classification. E2-MIL is mainly composed of three modules: a detail-aware attention distillation module (DAM), a structure-aware attention refined module (SRM), and an uncertainty-aware instance classifier (UIC). Specifically, DAM helps the global network locate more detail-aware positive instances by utilizing the complementary sub-bags to learn detailed attention knowledge from the local network. In addition, a masked self-guidance loss is also introduced to help bridge the gap between the slide-level labels and instance-level classification tasks. SRM generates a structure-aware attention map that locates the entire tumor region structure by effectively modeling the spatial relations between clustering instances. Moreover, UIC provides accurate instance-level classification results and robust predictive uncertainty estimation to improve the model reliability based on subjective logic theory. Extensive experiments on three large multi-center subtyping datasets demonstrate both slide-level and instance-level performance superiority of E2-MIL.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    基于组织病理学图像的生存预测旨在提供癌症预后的精确评估,并且可以告知个性化的治疗决策,以改善患者的预后。然而,现有方法无法自动对每个整张幻灯片图像(WSI)中的许多形态多样的小块之间的复杂相关性进行建模,从而阻止他们对患者状况有更深刻的理解和推断。为了解决这个问题,在这里,我们提出了一个新的深度学习框架,称为双流多依赖图神经网络(DM-GNN),以实现精确的癌症患者生存分析。具体来说,DM-GNN具有特征更新和全局分析分支,可以基于形态亲和力和全局共激活依赖性将每个WSI更好地建模为两个图。由于这两个依赖性从不同但互补的角度描绘了每个WSI,DM-GNN的两个设计分支可以共同实现补丁之间复杂相关性的多视图建模。此外,DM-GNN还能够通过引入亲和性引导注意力重新校准模块作为读出功能来提高图形构造期间依赖性信息的利用。这个新颖的模块提供了对特征扰动的增强的鲁棒性,从而确保更可靠和稳定的预测。在五个TCGA数据集上进行的广泛基准测试实验表明,DM-GNN优于其他最先进的方法,并基于高注意力补丁的形态学描述提供了可解释的预测见解。总的来说,DM-GNN代表了从组织病理学图像中个性化癌症预后的强大辅助工具,并且具有帮助临床医生做出个性化治疗决策和改善患者预后的巨大潜力。
    Histopathology image-based survival prediction aims to provide a precise assessment of cancer prognosis and can inform personalized treatment decision-making in order to improve patient outcomes. However, existing methods cannot automatically model the complex correlations between numerous morphologically diverse patches in each whole slide image (WSI), thereby preventing them from achieving a more profound understanding and inference of the patient status. To address this, here we propose a novel deep learning framework, termed dual-stream multi-dependency graph neural network (DM-GNN), to enable precise cancer patient survival analysis. Specifically, DM-GNN is structured with the feature updating and global analysis branches to better model each WSI as two graphs based on morphological affinity and global co-activating dependencies. As these two dependencies depict each WSI from distinct but complementary perspectives, the two designed branches of DM-GNN can jointly achieve the multi-view modeling of complex correlations between the patches. Moreover, DM-GNN is also capable of boosting the utilization of dependency information during graph construction by introducing the affinity-guided attention recalibration module as the readout function. This novel module offers increased robustness against feature perturbation, thereby ensuring more reliable and stable predictions. Extensive benchmarking experiments on five TCGA datasets demonstrate that DM-GNN outperforms other state-of-the-art methods and offers interpretable prediction insights based on the morphological depiction of high-attention patches. Overall, DM-GNN represents a powerful and auxiliary tool for personalized cancer prognosis from histopathology images and has great potential to assist clinicians in making personalized treatment decisions and improving patient outcomes.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    背景:数字整片图像(WSI)的出现推动了计算病理学的发展。然而,由于WSI的高分辨率,获得补丁级别的注释是具有挑战性和耗时的,这限制了完全监督方法的适用性。我们的目标是解决与补丁级别注释相关的挑战。
    方法:我们提出了一种基于多实例学习(MIL)的弱监督WSI分析的通用框架。为了实现实例功能的有效聚合,我们通过考虑特征分布,从多个维度设计了一个特征聚合模块,实例相关性和实例级评估。首先,我们实现了实例级标准化层和深度投影单元,以改善特征空间中实例的分离。然后,采用自我注意机制来探索实例之间的依赖关系。此外,引入了一种实例级伪标签评估方法,以增强弱监督过程中的可用信息。最后,使用袋级分类器来获得初步的WSI分类结果。为了实现更准确的WSI标签预测,我们设计了一个关键实例选择模块,加强了实例本地特征的学习。组合来自两个模块的结果导致WSI预测准确性的提高。
    结果:对Camelyon16,TCGA-NSCLC,SICAPv2,PANDA和经典的MIL基准测试数据集表明,与一些最近的方法相比,我们提出的方法具有竞争力。在分类精度方面最大提高14.6%。
    结论:我们的方法可以以弱监督的方式提高整个幻灯片图像的分类精度,更准确地检测病变区域。
    BACKGROUND: The emergence of digital whole slide image (WSI) has driven the development of computational pathology. However, obtaining patch-level annotations is challenging and time-consuming due to the high resolution of WSI, which limits the applicability of fully supervised methods. We aim to address the challenges related to patch-level annotations.
    METHODS: We propose a universal framework for weakly supervised WSI analysis based on Multiple Instance Learning (MIL). To achieve effective aggregation of instance features, we design a feature aggregation module from multiple dimensions by considering feature distribution, instances correlation and instance-level evaluation. First, we implement instance-level standardization layer and deep projection unit to improve the separation of instances in the feature space. Then, a self-attention mechanism is employed to explore dependencies between instances. Additionally, an instance-level pseudo-label evaluation method is introduced to enhance the available information during the weak supervision process. Finally, a bag-level classifier is used to obtain preliminary WSI classification results. To achieve even more accurate WSI label predictions, we have designed a key instance selection module that strengthens the learning of local features for instances. Combining the results from both modules leads to an improvement in WSI prediction accuracy.
    RESULTS: Experiments conducted on Camelyon16, TCGA-NSCLC, SICAPv2, PANDA and classical MIL benchmark datasets demonstrate that our proposed method achieves a competitive performance compared to some recent methods, with maximum improvement of 14.6 % in terms of classification accuracy.
    CONCLUSIONS: Our method can improve the classification accuracy of whole slide images in a weakly supervised way, and more accurately detect lesion areas.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    整个幻灯片图像(WSI)分析,由深度学习算法驱动,有可能彻底改变肿瘤检测,分类,和治疗反应预测。然而,挑战依然存在,例如在各种癌症类型中有限的模型泛化性,补丁级注释的劳动密集型性质,以及整合多倍放大信息以全面了解病理模式的必要性。
    为了应对这些挑战,我们介绍MAMILNet,用于WSI分析的创新多尺度注意多实例学习框架。将注意力机制纳入MAMILNet有助于其在不同癌症类型和预测任务中的非凡普遍性。此模型将整个幻灯片视为“bags”,将单个修补程序视为“实例”。“通过采用这种方法,MAMILNet有效地消除了复杂的补丁级别标签的要求,显著减少病理学家的人工工作量。为了提高预测准确性,该模型采用了多尺度的“咨询”策略,促进从各种放大倍数汇总测试结果。
    我们对MAMILNet的评估包括1171例,包括多种癌症类型,展示其在预测复杂任务方面的有效性。值得注意的是,MAMILNet在不同领域取得了令人印象深刻的成果:用于乳腺癌肿瘤检测,曲线下面积(AUC)为0.8872,准确度为0.8760。在肺癌分型诊断领域,它实现了0.9551的AUC和0.9095的准确性。此外,在预测卵巢癌的药物治疗反应方面,MAMILNet实现了0.7358的AUC和0.7341的准确性。
    这项研究的结果强调了MAMILNet在推动肿瘤领域内精准医学和个性化治疗计划的发展方面的潜力。通过有效解决与模型泛化相关的挑战,注释工作负载,和多放大倍数集成,MAMILNet在增强癌症患者的医疗保健结果方面显示出希望。该框架在准确检测乳腺肿瘤方面的成功,诊断肺癌类型,预测卵巢癌治疗反应突出了其对该领域的重大贡献,并为改善患者护理铺平了道路。
    UNASSIGNED: Whole Slide Image (WSI) analysis, driven by deep learning algorithms, has the potential to revolutionize tumor detection, classification, and treatment response prediction. However, challenges persist, such as limited model generalizability across various cancer types, the labor-intensive nature of patch-level annotation, and the necessity of integrating multi-magnification information to attain a comprehensive understanding of pathological patterns.
    UNASSIGNED: In response to these challenges, we introduce MAMILNet, an innovative multi-scale attentional multi-instance learning framework for WSI analysis. The incorporation of attention mechanisms into MAMILNet contributes to its exceptional generalizability across diverse cancer types and prediction tasks. This model considers whole slides as \"bags\" and individual patches as \"instances.\" By adopting this approach, MAMILNet effectively eliminates the requirement for intricate patch-level labeling, significantly reducing the manual workload for pathologists. To enhance prediction accuracy, the model employs a multi-scale \"consultation\" strategy, facilitating the aggregation of test outcomes from various magnifications.
    UNASSIGNED: Our assessment of MAMILNet encompasses 1171 cases encompassing a wide range of cancer types, showcasing its effectiveness in predicting complex tasks. Remarkably, MAMILNet achieved impressive results in distinct domains: for breast cancer tumor detection, the Area Under the Curve (AUC) was 0.8872, with an Accuracy of 0.8760. In the realm of lung cancer typing diagnosis, it achieved an AUC of 0.9551 and an Accuracy of 0.9095. Furthermore, in predicting drug therapy responses for ovarian cancer, MAMILNet achieved an AUC of 0.7358 and an Accuracy of 0.7341.
    UNASSIGNED: The outcomes of this study underscore the potential of MAMILNet in driving the advancement of precision medicine and individualized treatment planning within the field of oncology. By effectively addressing challenges related to model generalization, annotation workload, and multi-magnification integration, MAMILNet shows promise in enhancing healthcare outcomes for cancer patients. The framework\'s success in accurately detecting breast tumors, diagnosing lung cancer types, and predicting ovarian cancer therapy responses highlights its significant contribution to the field and paves the way for improved patient care.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    目的:病理图像分类是癌症诊断中最重要的辅助过程之一。为了克服标签弱的全幻灯片图像(WSI)样本不足的问题,基于伪包的多实例学习(MIL)方法在病理图像分类中引起了广泛的关注。在这种类型的方法中,伪袋的划分方案通常是影响分类性能的主要因素。为了改进现有随机/聚类方法上WSI伪包的划分,本文提出了一种新的原型驱动划分(ProDiv)方案,用于基于伪袋的病理图像MIL分类框架。
    方法:该方案首先设计了一种基于注意力的方法,为每张幻灯片生成一个袋子原型。在此基础上,它进一步根据原型和补丁之间的特征相似性将WSI补丁实例分为一系列实例集群。最后,伪包是通过随机组合不同实例集群的非重叠补丁实例得到的。此外,我们ProDiv的设计方案考虑了实用性,并且可以与近年来几乎所有基于MIL的WSI分类方法顺利组装。
    结果:实证结果表明,我们的ProDiv,当与几种现有方法集成时,可以提供高达7.3%和10.3%的分类AUC改进,分别在两个公共WSI数据集上。
    结论:与典型指标上的MIL模型相比,ProDiv几乎总是可以带来明显的性能改进,这表明了我们计划的有效性.实验可视化还可以直观地解释所提出的ProDiv的正确性。
    OBJECTIVE: Pathology image classification is one of the most essential auxiliary processes in cancer diagnosis. To overcome the problem of inadequate Whole-Slide Image (WSI) samples with weak labels, pseudo-bag-based multiple instance learning (MIL) methods have attracted wide attention in pathology image classification. In this type of method, the division scheme of pseudo-bags is usually a primary factor affecting classification performance. In order to improve the division of WSI pseudo-bags on existing random/clustering approaches, this paper proposes a new Prototype-driven Division (ProDiv) scheme for the pseudo-bag-based MIL classification framework on pathology images.
    METHODS: This scheme first designs an attention-based method to generate a bag prototype for each slide. On this basis, it further groups WSI patch instances into a series of instance clusters according to the feature similarities between the prototype and patches. Finally, pseudo-bags are obtained by randomly combining the non-overlapping patch instances of different instance clusters. Moreover, the design scheme of our ProDiv considers practicality, and it could be smoothly assembled with almost all the MIL-based WSI classification methods in recent years.
    RESULTS: Empirical results show that our ProDiv, when integrated with several existing methods, can deliver classification AUC improvements of up to 7.3% and 10.3%, respectively on two public WSI datasets.
    CONCLUSIONS: ProDiv could almost always bring obvious performance improvements to compared MIL models on typical metrics, which suggests the effectiveness of our scheme. Experimental visualization also visually interprets the correctness of the proposed ProDiv.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    人工智能驱动的计算病理学显著提高了肿瘤诊断的速度和精度,同时也显示出推断基因突变和基因表达水平的巨大潜力。然而,目前的研究在预测乳腺癌的分子亚型和临床结局方面仍然有限.在本文中,我们提出了一个弱监督对比学习框架来应对这一挑战。我们的框架首先对从整个幻灯片图像(WSI)平铺的大量未标记的补丁进行了对比学习预训练,以提取补丁级别的特征。利用门控注意力机制来聚合补丁级别的功能,以产生幻灯片功能,然后将其应用于各种下游任务。为了证实该方法的有效性,已使用3个公共队列和1个外部独立的乳腺癌队列进行评估实验.我们的模型推断基因表达的预测能力,分子亚型,在所有队列中验证了复发事件和药物反应.此外,学习到的斑块级注意力得分使我们能够生成与病理学家注释和空间转录组数据高度一致的热图.这些结果表明,我们的模型有效地建立了高阶基因型-表型关联,从而有可能扩展数字病理学在临床实践中的应用。
    The artificial intelligence-powered computational pathology has led to significant improvements in the speed and precision of tumor diagnosis, while also exhibiting substantial potential to infer genetic mutations and gene expression levels. However, current studies remain limited in predicting molecular subtypes and clinical outcomes in breast cancer. In this paper, we proposed a weakly supervised contrastive learning framework to address this challenge. Our framework first performed contrastive learning pretraining on a large number of unlabeled patches tiled from whole slide images (WSIs) to extract patch-level features. The gated attention mechanism was leveraged to aggregate patch-level features to produce slide feature that was then applied to various downstream tasks. To confirm the effectiveness of the proposed method, three public cohorts and one external independent cohort of breast cancer have been used to conducted evaluation experiments. The predictive powers of our model to infer gene expression, molecular subtypes, recurrence events and drug responses were validated across cohorts. In addition, the learned patch-level attention scores enabled us to generate heatmaps that were highly consistent with pathologist annotations and spatial transcriptomic data. These findings demonstrated that our model effectively established the high-order genotype-phenotype associations, thereby potentially extend the application of digital pathology in clinical practice.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    文献中的弱监督图像分割方法通常使用紧密的边界框监督来实现高分割性能,而在由松散的边界框监督时,性能会大大降低。然而,与松散的边界框相比,由于对框的四个侧面的精确位置有严格的要求,因此获得紧密的边界框更加困难。要解决此问题,本研究调查是否有可能保持良好的分割性能时,松散的边界框被用作监督。为此,这项工作通过集成基于极坐标变换的MIL策略来辅助图像分割,扩展了我们以前基于并行变换的多实例学习(MIL)的紧密边界框监督。提出的基于极坐标变换的MIL公式适用于紧密和松散的边界框,其中,正包定义为边界框的极坐标线中的像素,其中一个端点位于由框包围的对象内部,另一个端点位于框的四个边之一。此外,引入了加权平滑最大近似,以结合观察到更靠近极坐标变换原点的像素更有可能属于框中的对象。当实验中考虑了不同精度水平的边界框时,使用骰子系数在两个公共数据集上对所提出的方法进行了评估。结果表明,所提出的方法在所有精度水平上都实现了边界框的最新性能,并且对松散边界框注释中的轻度和中度错误具有鲁棒性。这些代码可在https://github.com/wangjuan313/wsis-beyond-tightBB上获得。
    Weakly supervised image segmentation approaches in the literature usually achieve high segmentation performance using tight bounding box supervision and decrease the performance greatly when supervised by loose bounding boxes. However, compared with loose bounding box, it is much more difficult to acquire tight bounding box due to its strict requirements on the precise locations of the four sides of the box. To resolve this issue, this study investigates whether it is possible to maintain good segmentation performance when loose bounding boxes are used as supervision. For this purpose, this work extends our previous parallel transformation based multiple instance learning (MIL) for tight bounding box supervision by integrating an MIL strategy based on polar transformation to assist image segmentation. The proposed polar transformation based MIL formulation works for both tight and loose bounding boxes, in which a positive bag is defined as pixels in a polar line of a bounding box with one endpoint located inside the object enclosed by the box and the other endpoint located at one of the four sides of the box. Moreover, a weighted smooth maximum approximation is introduced to incorporate the observation that pixels closer to the origin of the polar transformation are more likely to belong to the object in the box. The proposed approach was evaluated on two public datasets using dice coefficient when bounding boxes at different precision levels were considered in the experiments. The results demonstrate that the proposed approach achieves state-of-the-art performance for bounding boxes at all precision levels and is robust to mild and moderate errors in the loose bounding box annotations. The codes are available at https://github.com/wangjuan313/wsis-beyond-tightBB.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

公众号