Self-supervised learning

自监督学习
  • 文章类型: Journal Article
    深度神经网络在监督学习任务中表现出色,但由于需要大量标记数据而受到限制。自我监督学习成为一种有前途的选择,允许模型在没有明确标签的情况下学习。信息论塑造了深度神经网络,特别是信息瓶颈原则。该原理优化了压缩和保留相关信息之间的权衡,为监督环境中的有效网络设计提供基础。然而,其在自监督学习中的确切作用和适应性尚不清楚。在这项工作中,我们从信息理论的角度审视各种自监督学习方法,引入一个统一的框架来封装自监督信息理论学习问题。这个框架包括多个编码器和解码器,这表明,所有现有的自监督学习工作都可以看作是具体的例子。我们的目标是统一这些方法,以更好地理解它们的基本原理,并解决主要挑战:许多作品提出了不同的框架和不同的理论,这似乎是矛盾的。通过将现有的研究编织成一个有凝聚力的叙事,我们深入研究当代的自我监督方法,聚光灯潜在的研究领域,并强调固有的挑战。此外,我们讨论了如何估计信息论量及其相关的经验问题。总的来说,本文对信息论的交叉进行了全面的回顾,自我监督学习,和深度神经网络,旨在通过我们提出的统一方法更好地理解。
    Deep neural networks excel in supervised learning tasks but are constrained by the need for extensive labeled data. Self-supervised learning emerges as a promising alternative, allowing models to learn without explicit labels. Information theory has shaped deep neural networks, particularly the information bottleneck principle. This principle optimizes the trade-off between compression and preserving relevant information, providing a foundation for efficient network design in supervised contexts. However, its precise role and adaptation in self-supervised learning remain unclear. In this work, we scrutinize various self-supervised learning approaches from an information-theoretic perspective, introducing a unified framework that encapsulates the self-supervised information-theoretic learning problem. This framework includes multiple encoders and decoders, suggesting that all existing work on self-supervised learning can be seen as specific instances. We aim to unify these approaches to understand their underlying principles better and address the main challenge: many works present different frameworks with differing theories that may seem contradictory. By weaving existing research into a cohesive narrative, we delve into contemporary self-supervised methodologies, spotlight potential research areas, and highlight inherent challenges. Moreover, we discuss how to estimate information-theoretic quantities and their associated empirical problems. Overall, this paper provides a comprehensive review of the intersection of information theory, self-supervised learning, and deep neural networks, aiming for a better understanding through our proposed unified approach.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Systematic Review
    医疗时间序列是随着时间的推移收集的顺序数据,用于衡量与健康相关的信号,如脑电图(EEG),心电图(ECG),和重症监护病房(ICU)的读数。分析医疗时间序列并确定潜在的模式和趋势,从而发现高度有价值的见解,以增强诊断,治疗,风险评估,和疾病进展。然而,医学时间序列中的数据挖掘受到耗时耗力的样本标注的严重限制,和专家依赖。为了缓解这一挑战,新兴的自监督对比学习,自2020年以来取得了巨大的成功,是一个有前途的解决方案。对比学习旨在通过对比正样本和负样本来学习代表性嵌入,而无需显式标签。这里,我们根据PRISMA标准,对对比学习如何缓解医学时间序列中的标签缺乏进行了系统评价.我们在五个科学数据库(IEEE,ACM,Scopus,谷歌学者,和PubMed),并根据纳入标准检索了1908年的论文。应用排除标准后,和标题筛选,abstract,和全文水平,我们仔细审查了这方面的43篇论文。具体来说,本文概述了对比学习的流水线,包括训练前,微调,和测试。我们提供了应用于医疗时间序列数据的各种增强的全面总结,预训练编码器的架构,微调分类器和聚类的类型,和流行的对比损失函数。此外,我们概述了医疗时间序列中使用的不同数据类型,突出感兴趣的医学应用,并提供已在该领域使用的51个公共数据集的综合表格。此外,本文将对未来有前途的范围进行讨论,例如为有效的增强设计提供指导,开发一个分析分层时间序列的统一框架,和研究处理多模态数据的方法。尽管处于早期阶段,自我监督对比学习在克服医学时间序列研究中对专家创建注释的需求方面显示出巨大的潜力。
    Medical time series are sequential data collected over time that measures health-related signals, such as electroencephalography (EEG), electrocardiography (ECG), and intensive care unit (ICU) readings. Analyzing medical time series and identifying the latent patterns and trends that lead to uncovering highly valuable insights for enhancing diagnosis, treatment, risk assessment, and disease progression. However, data mining in medical time series is heavily limited by the sample annotation which is time-consuming and labor-intensive, and expert-depending. To mitigate this challenge, the emerging self-supervised contrastive learning, which has shown great success since 2020, is a promising solution. Contrastive learning aims to learn representative embeddings by contrasting positive and negative samples without the requirement for explicit labels. Here, we conducted a systematic review of how contrastive learning alleviates the label scarcity in medical time series based on PRISMA standards. We searched the studies in five scientific databases (IEEE, ACM, Scopus, Google Scholar, and PubMed) and retrieved 1908 papers based on the inclusion criteria. After applying excluding criteria, and screening at title, abstract, and full text levels, we carefully reviewed 43 papers in this area. Specifically, this paper outlines the pipeline of contrastive learning, including pre-training, fine-tuning, and testing. We provide a comprehensive summary of the various augmentations applied to medical time series data, the architectures of pre-training encoders, the types of fine-tuning classifiers and clusters, and the popular contrastive loss functions. Moreover, we present an overview of the different data types used in medical time series, highlight the medical applications of interest, and provide a comprehensive table of 51 public datasets that have been utilized in this field. In addition, this paper will provide a discussion on the promising future scopes such as providing guidance for effective augmentation design, developing a unified framework for analyzing hierarchical time series, and investigating methods for processing multimodal data. Despite being in its early stages, self-supervised contrastive learning has shown great potential in overcoming the need for expert-created annotations in the research of medical time series.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

公众号