Weakly supervised learning

弱监督学习
  • 文章类型: Journal Article
    对不同阶段的皮肤癌病灶进行精确分割,有利于早期发现和进一步治疗。考虑到为这项任务获得像素完美注释的巨大成本,使用成本较低的图像级标签进行分割已成为一个研究方向。大多数图像级标签弱监督分割使用类激活映射(CAM)方法。这种方法的常见结果是前景分割不完整,分割不足,或者假阴性。同时,在对皮肤癌病变进行弱监督分割时,溃疡,发红,和肿胀可能出现在个别疾病类别的分割区域附近。这种共存问题在一定程度上影响了模型分割类相关组织边界的准确性。以上两个问题是由惩罚整个图像空间的图像级标签的松散约束性质决定的。因此,为图像级标签的弱监督提供像素级约束是提高性能的关键。为了解决上述问题,提出了一种联合无监督约束辅助弱监督分割模型(UCA-WSS)。模型的弱监督部分采用双分支对抗擦除机制来生成更高质量的CAM。无监督部分采用对比学习和聚类算法生成前景标签和精细边界标签,通过无监督约束辅助分割,解决弱监督皮肤癌病灶分割中常见的共现问题。本文提出的模型在一些公共皮肤病学数据集上与其他相关模型进行了比较评估。实验结果表明,与其他弱监督分割模型相比,我们的模型在皮肤癌分割任务上的表现更好。显示了在弱监督分割上结合无监督约束方法的潜力。
    Precise segmentation for skin cancer lesions at different stages is conducive to early detection and further treatment. Considering the huge cost of obtaining pixel-perfect annotations for this task, segmentation using less expensive image-level labels has become a research direction. Most image-level label weakly supervised segmentation uses class activation mapping (CAM) methods. A common consequence of this method is incomplete foreground segmentation, insufficient segmentation, or false negatives. At the same time, when performing weakly supervised segmentation of skin cancer lesions, ulcers, redness, and swelling may appear near the segmented areas of individual disease categories. This co-occurrence problem affects the model\'s accuracy in segmenting class-related tissue boundaries to a certain extent. The above two issues are determined by the loosely constrained nature of image-level labels that penalize the entire image space. Therefore, providing pixel-level constraints for weak supervision of image-level labels is the key to improving performance. To solve the above problems, this paper proposes a joint unsupervised constraint-assisted weakly supervised segmentation model (UCA-WSS). The weakly supervised part of the model adopts a dual-branch adversarial erasure mechanism to generate higher-quality CAM. The unsupervised part uses contrastive learning and clustering algorithms to generate foreground labels and fine boundary labels to assist segmentation and solve common co-occurrence problems in weakly supervised skin cancer lesion segmentation through unsupervised constraints. The model proposed in the article is evaluated comparatively with other related models on some public dermatology data sets. Experimental results show that our model performs better on the skin cancer segmentation task than other weakly supervised segmentation models, showing the potential of combining unsupervised constraint methods on weakly supervised segmentation.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    尽管深度学习已经实现了自动医学图像分割的最先进性能,它通常需要大量的像素级手动注释来进行训练。获得这些高质量的注释非常耗时,并且需要专业知识,这阻碍了依赖于此类注释来训练具有良好分割性能的模型的广泛应用。使用涂鸦注释可以大大降低注释成本,但由于监管不足,往往会导致细分性能不佳。在这项工作中,我们提出了一个名为ScribSD+的新框架,该框架基于多尺度知识蒸馏和类式对比正则化,用于从涂鸦注释中学习。对于由涂鸦和基于指数移动平均(EMA)的教师监督的学生网络,我们首先介绍了多尺度预测水平知识蒸馏(KD),它利用教师网络的软预测在多个尺度上监督学生,然后提出类对比正则化,鼓励同一类内的特征相似性和不同类之间的差异,从而有效提高学生网络的分割性能。用于心脏结构分割的ACDC数据集和用于胎盘和胎儿脑分割的胎儿MRI数据集上的实验结果表明,我们的方法显着提高了学生的表现,并且优于五种最先进的涂鸦监督学习方法。因此,该方法有可能降低开发用于临床诊断的深度学习模型的注释成本。
    Despite that deep learning has achieved state-of-the-art performance for automatic medical image segmentation, it often requires a large amount of pixel-level manual annotations for training. Obtaining these high-quality annotations is time-consuming and requires specialized knowledge, which hinders the widespread application that relies on such annotations to train a model with good segmentation performance. Using scribble annotations can substantially reduce the annotation cost, but often leads to poor segmentation performance due to insufficient supervision. In this work, we propose a novel framework named as ScribSD+ that is based on multi-scale knowledge distillation and class-wise contrastive regularization for learning from scribble annotations. For a student network supervised by scribbles and the teacher based on Exponential Moving Average (EMA), we first introduce multi-scale prediction-level Knowledge Distillation (KD) that leverages soft predictions of the teacher network to supervise the student at multiple scales, and then propose class-wise contrastive regularization which encourages feature similarity within the same class and dissimilarity across different classes, thereby effectively improving the segmentation performance of the student network. Experimental results on the ACDC dataset for heart structure segmentation and a fetal MRI dataset for placenta and fetal brain segmentation demonstrate that our method significantly improves the student\'s performance and outperforms five state-of-the-art scribble-supervised learning methods. Consequently, the method has a potential for reducing the annotation cost in developing deep learning models for clinical diagnosis.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    乳腺癌的分子分类对于有效治疗至关重要。数字病理学的出现开创了一个新时代,在这个时代中,利用整个幻灯片图像的弱监督学习在开发深度学习模型中获得了突出地位,因为这种方法减轻了对大量手动注释的需求。采用弱监督学习对乳腺癌分子亚型进行分类。
    我们的方法利用了两个全幻灯片图像数据集:一个由来自韩国大学Guro医院(KG)的乳腺癌病例组成,另一个来自癌症基因组图谱数据集(TCGA)。此外,我们使用基于注意力的热图可视化了推断的结果,并回顾了最专注的斑块的组织形态学特征.
    KG+TCGA训练的模型实现了0.749的接收器操作特征值下的区域。一个固有的挑战在于亚型之间的不平衡。此外,两个数据集之间的差异导致不同的分子亚型比例。为了缓解这种不平衡,我们合并了两个数据集,所得到的模型表现出改进的性能。注意的斑块与广泛认可的组织形态学特征密切相关。三阴性亚型高等级核的发生率高,肿瘤坏死,和肿瘤内浸润淋巴细胞。腔A亚型显示胶原纤维的高发生率。
    基于弱监督学习的人工智能(AI)模型显示出有希望的性能。对最专注的补丁的回顾提供了对AI模型预测的见解。人工智能模型可以成为宝贵的筛选工具,在实践中降低成本和工作量。
    UNASSIGNED: The molecular classification of breast cancer is crucial for effective treatment. The emergence of digital pathology has ushered in a new era in which weakly supervised learning leveraging whole-slide images has gained prominence in developing deep learning models because this approach alleviates the need for extensive manual annotation. Weakly supervised learning was employed to classify the molecular subtypes of breast cancer.
    UNASSIGNED: Our approach capitalizes on two whole-slide image datasets: one consisting of breast cancer cases from the Korea University Guro Hospital (KG) and the other originating from The Cancer Genomic Atlas dataset (TCGA). Furthermore, we visualized the inferred results using an attention-based heat map and reviewed the histomorphological features of the most attentive patches.
    UNASSIGNED: The KG+TCGA-trained model achieved an area under the receiver operating characteristics value of 0.749. An inherent challenge lies in the imbalance among subtypes. Additionally, discrepancies between the two datasets resulted in different molecular subtype proportions. To mitigate this imbalance, we merged the two datasets, and the resulting model exhibited improved performance. The attentive patches correlated well with widely recognized histomorphologic features. The triple-negative subtype has a high incidence of high-grade nuclei, tumor necrosis, and intratumoral tumor-infiltrating lymphocytes. The luminal A subtype showed a high incidence of collagen fibers.
    UNASSIGNED: The artificial intelligence (AI) model based on weakly supervised learning showed promising performance. A review of the most attentive patches provided insights into the predictions of the AI model. AI models can become invaluable screening tools that reduce costs and workloads in practice.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    磁共振成像(MRI)图像的注释在基于深度学习的MRI分割任务中起着重要作用。半自动标注算法有助于提高MRI图像标注的效率,降低图像标注的难度。然而,现有的基于深度学习的半自动标注算法在分割标签不足的情况下预标注性能较差。在本文中,提出了一种基于半弱监督学习的半自动MRI标注算法。为了在分割标签不足的情况下实现更好的预标注性能,引入了半监督学习和弱监督学习,提出了一种基于稀疏标签的半弱监督学习分割算法。此外,为了提高单个分割标签对预标注模型性能的贡献率,设计了一种基于主动学习的迭代标注策略。在公开的MRI数据集上的实验结果表明,当分割标签数量远少于完全监督学习算法时,该算法实现了等效的预标注性能。证明了该算法的有效性。
    The annotation of magnetic resonance imaging (MRI) images plays an important role in deep learning-based MRI segmentation tasks. Semi-automatic annotation algorithms are helpful for improving the efficiency and reducing the difficulty of MRI image annotation. However, the existing semi-automatic annotation algorithms based on deep learning have poor pre-annotation performance in the case of insufficient segmentation labels. In this paper, we propose a semi-automatic MRI annotation algorithm based on semi-weakly supervised learning. In order to achieve a better pre-annotation performance in the case of insufficient segmentation labels, semi-supervised and weakly supervised learning were introduced, and a semi-weakly supervised learning segmentation algorithm based on sparse labels was proposed. In addition, in order to improve the contribution rate of a single segmentation label to the performance of the pre-annotation model, an iterative annotation strategy based on active learning was designed. The experimental results on public MRI datasets show that the proposed algorithm achieved an equivalent pre-annotation performance when the number of segmentation labels was much less than that of the fully supervised learning algorithm, which proves the effectiveness of the proposed algorithm.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    间质性肺病(ILD)的特征是进行性病理变化,需要及时准确的诊断。ILD的早期发现和进展评估对于有效管理很重要。这项研究引入了一种新颖的定量评估方法,该方法利用胸部X光片分析ILD的像素级变化。使用弱监督学习框架,该方法结合了对比非配对翻译模型和新开发的ILD范围评分算法,与传统的视觉评估相比,可以更精确和客观地量化疾病变化.通过该方法计算的ILD程度分数表明ILD和正常类别之间的分类准确度为92.98%。此外,使用ILD随访数据集进行间隔变化分析,该方法评估疾病进展的准确率为85.29%.这些发现验证了ILD程度评分作为ILD监测工具的可靠性。这项研究的结果表明,提出的定量方法可以改善ILD的监测和管理。
    Interstitial lung disease (ILD) is characterized by progressive pathological changes that require timely and accurate diagnosis. The early detection and progression assessment of ILD are important for effective management. This study introduces a novel quantitative evaluation method utilizing chest radiographs to analyze pixel-wise changes in ILD. Using a weakly supervised learning framework, the approach incorporates the contrastive unpaired translation model and a newly developed ILD extent scoring algorithm for more precise and objective quantification of disease changes than conventional visual assessments. The ILD extent score calculated through this method demonstrated a classification accuracy of 92.98% between ILD and normal classes. Additionally, using an ILD follow-up dataset for interval change analysis, this method assessed disease progression with an accuracy of 85.29%. These findings validate the reliability of the ILD extent score as a tool for ILD monitoring. The results of this study suggest that the proposed quantitative method may improve the monitoring and management of ILD.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    仪器位姿估计是计算机辅助手术的关键需求,其主要挑战在于两个方面:第一,由于仪器的高折射和复杂的背景,难以获得稳定的相应图像特征点,其次,缺乏标记的姿势数据。本研究旨在使用单个内窥镜图像解决当前内窥镜系统中手术器械的姿态估计问题。更具体地说,提出了一种基于仪器图像分割轮廓的弱监督方法,在合成内窥镜图像的有效辅助下。我们的方法由以下三个模块组成:一个分割模块,用于自动检测输入图像中的仪器,然后是点推断模块,以预测仪器的隐式特征点的图像位置,和点向后传播的透视n点模块,用于从试探性2D-3D对应点估计姿态。为了缓解对点对应准确性的过度依赖,特征点匹配的局部误差和相应轮廓的全局不一致性同时最小化。与当前最先进的方法相比,我们提出的方法在真实和合成图像上都得到了验证。
    Instrument pose estimation is a key demand in computer-aided surgery, and its main challenges lie in two aspects: Firstly, the difficulty of obtaining stable corresponding image feature points due to the instruments\' high refraction and complicated background, and secondly, the lack of labeled pose data. This study aims to tackle the pose estimation problem of surgical instruments in the current endoscope system using a single endoscopic image. More specifically, a weakly supervised method based on the instrument\'s image segmentation contour is proposed, with the effective assistance of synthesized endoscopic images. Our method consists of the following three modules: a segmentation module to automatically detect the instrument in the input image, followed by a point inference module to predict the image locations of the implicit feature points of the instrument, and a point back-propagatable Perspective-n-Point module to estimate the pose from the tentative 2D-3D corresponding points. To alleviate the over-reliance on point correspondence accuracy, the local errors of feature point matching and the global inconsistency of the corresponding contours are simultaneously minimized. Our proposed method is validated with both real and synthetic images in comparison with the current state-of-the-art methods.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    结直肠癌(CRC)是美国第三大最常见的癌症。肿瘤出芽(TB)检测和定量是通过组织病理学图像分析确定CRC阶段的关键但劳动密集型步骤。为了帮助这个过程,我们使用SAM-Adapter对CRC组织病理学图像上的任意段模型(SAM)进行调整以分割TB。在这种方法中,我们会自动从CRC图像中获取特定于任务的提示,并以参数有效的方式训练SAM模型。我们使用病理学家的注释将模型的预测与从零开始训练的模型的预测进行比较。因此,我们的模型实现了0.65的交集联合(IoU)和0.75的实例级Dice评分,这在匹配病理学家的TB注释方面是有希望的。我们相信我们的研究提供了一种新的解决方案来识别H&E染色的组织病理学图像上的TBs。我们的研究还证明了将基础模型用于病理图像分割任务的价值。
    Colorectal cancer (CRC) is the third most common cancer in the United States. Tumor Budding (TB) detection and quantification are crucial yet labor-intensive steps in determining the CRC stage through the analysis of histopathology images. To help with this process, we adapt the Segment Anything Model (SAM) on the CRC histopathology images to segment TBs using SAM-Adapter. In this approach, we automatically take task-specific prompts from CRC images and train the SAM model in a parameter-efficient way. We compare the predictions of our model with the predictions from a trained-from-scratch model using the annotations from a pathologist. As a result, our model achieves an intersection over union (IoU) of 0.65 and an instance-level Dice score of 0.75, which are promising in matching the pathologist\'s TB annotation. We believe our study offers a novel solution to identify TBs on H&E-stained histopathology images. Our study also demonstrates the value of adapting the foundation model for pathology image segmentation tasks.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    目的:当前的自动心电图(ECG)诊断系统可以提供分类结果,但通常缺乏对这些结果的解释。这种限制阻碍了它们在临床诊断中的应用。先前的监督学习无法在没有手动标记大型ECG数据集的情况下足够准确地突出异常分割输出以用于临床应用。
    方法:在本研究中,我们提出了一个名为MA-MIL的多实例学习框架,它设计了一个多层和多实例的结构,以不同的规模逐步聚合。我们使用公共MIT-BIH数据集和私有数据集评估了我们的方法。
    结果:结果表明,我们的模型在ECG分类输出和心跳水平方面均表现良好,次心跳级异常段检测,ECG分类的准确性和F1评分分别为0.987和0.986,心跳水平异常检测的准确性和F1评分分别为0.968和0.949,分别。与可视化方法相比,在所有类别中,MA-MIL的IoU值提高了至少17%,至多31%。
    结论:MA-MIL可以准确定位异常心电图段,为临床应用提供更可靠的结果。
    OBJECTIVE: Current automatic electrocardiogram (ECG) diagnostic systems could provide classification outcomes but often lack explanations for these results. This limitation hampers their application in clinical diagnoses. Previous supervised learning could not highlight abnormal segmentation output accurately enough for clinical application without manual labeling of large ECG datasets.
    METHODS: In this study, we present a multi-instance learning framework called MA-MIL, which has designed a multi-layer and multi-instance structure that is aggregated step by step at different scales. We evaluated our method using the public MIT-BIH dataset and our private dataset.
    RESULTS: The results show that our model performed well in both ECG classification output and heartbeat level, sub-heartbeat level abnormal segment detection, with accuracy and F1 scores of 0.987 and 0.986 for ECG classification and 0.968 and 0.949 for heartbeat level abnormal detection, respectively. Compared to visualization methods, the IoU values of MA-MIL improved by at least 17 % and at most 31 % across all categories.
    CONCLUSIONS: MA-MIL could accurately locate the abnormal ECG segment, offering more trustworthy results for clinical application.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    自动裂缝分割在维护建筑物和基础设施的结构健康方面起着至关重要的作用。尽管在完全监督的裂缝分割方面取得了成功,昂贵的像素级注释限制了其应用,导致弱监督裂纹分割(WSCS)的勘探增加。然而,WSCS方法不可避免地带来嘈杂的伪标签,这导致了巨大的波动。为了解决这个问题,我们提出了一种新的WSCS信心感知联合训练(CAC)框架。该框架旨在迭代地细化伪标签,促进学习更稳健的分割模型。具体来说,设计了一种协同训练机制,并构建了两个协作网络来学习不确定的裂纹像素,从容易到难。此外,动态划分策略设计为基于裂纹置信度分数划分伪标签。其中,高置信度伪标签用于优化协作网络的初始化参数,而低置信度伪标签丰富了裂纹样本的多样性。对Crack500,DeepCrack进行了广泛的实验,和CFD数据集表明,提出的CAC明显优于其他WSCS方法。
    Automatic crack segmentation plays an essential role in maintaining the structural health of buildings and infrastructure. Despite the success in fully supervised crack segmentation, the costly pixel-level annotation restricts its application, leading to increased exploration in weakly supervised crack segmentation (WSCS). However, WSCS methods inevitably bring in noisy pseudo-labels, which results in large fluctuations. To address this problem, we propose a novel confidence-aware co-training (CAC) framework for WSCS. This framework aims to iteratively refine pseudo-labels, facilitating the learning of a more robust segmentation model. Specifically, a co-training mechanism is designed and constructs two collaborative networks to learn uncertain crack pixels, from easy to hard. Moreover, the dynamic division strategy is designed to divide the pseudo-labels based on the crack confidence score. Among them, the high-confidence pseudo-labels are utilized to optimize the initialization parameters for the collaborative network, while low-confidence pseudo-labels enrich the diversity of crack samples. Extensive experiments conducted on the Crack500, DeepCrack, and CFD datasets demonstrate that the proposed CAC significantly outperforms other WSCS methods.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    弱监督时间动作定位旨在使用视频级别标签定位未修剪视频中动作实例的时间边界,并为其分配相应的动作类别。一般来说,它是通过一个名为“按分类本地化”的管道来解决的,,它通过对视频片段进行分类来查找动作实例。然而,由于这种方法优化了视频级分类目标,生成的激活序列经常受到类相关场景的干扰,导致预测结果出现大量误报。许多现有作品将背景视为一个独立的类别,迫使模型学习区分背景片段。然而,在弱监督条件下,背景信息是模糊和不确定的,使这种方法极其困难。为了减轻误报的影响,我们提出了一个新的行动导向的假阳性抑制框架。我们的方法试图在不引入背景类别的情况下抑制假阳性背景。首先,我们提出了一个自我训练行动分支来学习班级不可知的行动,可以通过忽略视频标签来最小化类相关场景信息的干扰。其次,我们提出了一个假阳性抑制模块来挖掘假阳性片段并对其进行抑制。最后,我们介绍了前景增强模块,它指导模型在注意力机制和类不可知动作的帮助下学习前景。我们对三个基准测试(THUMOS14、ActivityNet1.2和ActivityNet1.3)进行了广泛的实验。结果证明了我们的方法在抑制误报方面的有效性,并且达到了最先进的性能。代码:https://github.com/lizhilin-ustc/AFPS。
    Weakly supervised temporal action localization aims to locate the temporal boundaries of action instances in untrimmed videos using video-level labels and assign them the corresponding action category. Generally, it is solved by a pipeline called \"localization-by-classification\", which finds the action instances by classifying video snippets. However, since this approach optimizes the video-level classification objective, the generated activation sequences often suffer interference from class-related scenes, resulting in a large number of false positives in the prediction results. Many existing works treat background as an independent category, forcing models to learn to distinguish background snippets. However, under weakly supervised conditions, the background information is fuzzy and uncertain, making this method extremely difficult. To alleviate the impact of false positives, we propose a new actionness-guided false positive suppression framework. Our method seeks to suppress false positive backgrounds without introducing the background category. Firstly, we propose a self-training actionness branch to learn class-agnostic actionness, which can minimize the interference of class-related scene information by ignoring the video labels. Secondly, we propose a false positive suppression module to mine false positive snippets and suppress them. Finally, we introduce the foreground enhancement module, which guides the model to learn the foreground with the help of the attention mechanism as well as class-agnostic actionness. We conduct extensive experiments on three benchmarks (THUMOS14, ActivityNet1.2, and ActivityNet1.3). The results demonstrate the effectiveness of our method in suppressing false positives and it achieves the state-of-the-art performance. Code: https://github.com/lizhilin-ustc/AFPS.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

公众号