unsupervised learning

无监督学习
  • 文章类型: Journal Article
    磁共振成像(MRI)-计算机断层扫描(CT)合成在仅MRI放射治疗工作流程中至关重要,特别是通过以其准确性而闻名的深度学习技术。然而,当前的监督方法仅限于特定中心的学习,并依赖于配准精度。这项研究的目的是评估在前列腺MRI到CT生成放射治疗剂量计算的背景下,无监督和监督方法的准确性。
    使用来自三个不同中心的99名前列腺癌患者的CT/MRI图像对。对有监督和无监督条件生成对抗网络(cGAN)进行了比较。无监督训练将风格转移方法与。增强感知合成(CREP)损失的内容和风格表示。对于剂量评估,在体积调制电弧治疗(VMAT)中,光子处方剂量为60Gy.用于sCT评估的成像终点是平均绝对误差(MAE)。剂量学终点包括绝对剂量差异和CT和sCT剂量计算之间的伽马分析。
    无监督的配对网络在MAE为33.6HU时表现出最高的身体精度,通过无监督的非配对学习获得的最高MAE为45.5HU。所有架构提供了用于剂量计算的临床上可接受的结果,其中γ通过率高于94%(1%Imm10%)。
    这项研究表明,多中心数据可以通过无监督学习产生准确的sCT,消除CT-MRI配准。sCT不仅匹配HU值,而且能够进行精确的剂量计算,表明它们在仅MRI放射治疗工作流程中更广泛使用的潜力。
    UNASSIGNED: Magnetic resonance imaging (MRI)-to-computed tomography (CT) synthesis is essential in MRI-only radiotherapy workflows, particularly through deep learning techniques known for their accuracy. However, current supervised methods are limited to specific center\'s learnings and depend on registration precision. The aim of this study was to evaluate the accuracy of unsupervised and supervised approaches in the context of prostate MRI-to-CT generation for radiotherapy dose calculation.
    UNASSIGNED: CT/MRI image pairs from 99 prostate cancer patients across three different centers were used. A comparison between supervised and unsupervised conditional Generative Adversarial Networks (cGAN) was conducted. Unsupervised training incorporates a style transfer method with. Content and Style Representation for Enhanced Perceptual synthesis (CREPs) loss. For dose evaluation, the photon prescription dose was 60 Gy delivered in volumetric modulated arc therapy (VMAT). Imaging endpoint for sCT evaluation was Mean Absolute Error (MAE). Dosimetric endpoints included absolute dose differences and gamma analysis between CT and sCT dose calculations.
    UNASSIGNED: The unsupervised paired network exhibited the highest accuracy for the body with a MAE at 33.6 HU, the highest MAE was 45.5 HU obtained with unsupervised unpaired learning. All architectures provided clinically acceptable results for dose calculation with gamma pass rates above 94 % (1 % 1 mm 10 %).
    UNASSIGNED: This study shows that multicenter data can produce accurate sCTs via unsupervised learning, eliminating CT-MRI registration. The sCTs not only matched HU values but also enabled precise dose calculations, suggesting their potential for wider use in MRI-only radiotherapy workflows.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    目的:基于深度学习的图像增强在超声图像处理领域具有巨大的潜力,因为它可以准确地对复杂的非线性伪影和噪声进行建模,如超声波斑点图案。然而,训练深度学习网络以获取干净且无噪声的参考图像提出了重大挑战。本研究引入了一个无监督的深度学习框架,称为散斑到散斑(S2S),设计用于抑制斑点和噪声。该框架可以在不需要干净(无斑点)参考图像的情况下完成其训练。
    方法:所提出的网络利用统计推理来相互训练两个体内图像,每个都有不同的斑点图案和噪音。然后,它推断无斑点和噪声的图像,而不需要干净的参考图像。这种方法大大减少了时间,成本,和工作专家需要投资手动注释参考图像。
    结果:实验结果表明,所提出的方法在信噪比方面优于现有技术,对比噪声比,结构相似性指数,边缘保存指数,和处理时间(快86倍)。除了这项工作中使用的图像外,它还对从超声扫描仪获得的图像进行了出色的处理。
    结论:S2S证明了在医学成像应用中采用无监督学习技术的潜力,获取地面真相参考是具有挑战性的。
    OBJECTIVE: Deep learning-based image enhancement has significant potential in the field of ultrasound image processing, as it can accurately model complicated nonlinear artifacts and noise, such as ultrasonic speckle patterns. However, training deep learning networks to acquire reference images that are clean and free of noise presents significant challenges. This study introduces an unsupervised deep learning framework, termed speckle-to-speckle (S2S), designed for speckle and noise suppression. This framework can complete its training without the need for clean (speckle-free) reference images.
    METHODS: The proposed network leverages statistical reasoning for the mutual training of two in vivo images, each with distinct speckle patterns and noise. It then infers speckle- and noise-free images without needing clean reference images. This approach significantly reduces the time, cost, and effort experts need to invest in annotating reference images manually.
    RESULTS: The experimental results demonstrated that the proposed approach outperformed existing techniques in terms of the signal-to-noise ratio, contrast-to-noise ratio, structural similarity index, edge preservation index, and processing time (up to 86 times faster). It also performed excellently on images obtained from ultrasound scanners other than the ones used in this work.
    CONCLUSIONS: S2S demonstrates the potential of employing an unsupervised learning-based technique in medical imaging applications, where acquiring a ground truth reference is challenging.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    目的:光学相干断层扫描(OCT)因其无创性,高分辨率成像能力。然而,其低相干性原理固有的斑点噪声会降低图像质量并损害诊断准确性。虽然深度学习方法在减少斑点噪声方面表现出了希望,获得良好配准的图像对仍然具有挑战性,导致不成对方法的发展。尽管有潜力,现有的不成对方法遭受网络结构或交互机制的冗余。因此,非配对OCT去噪的更简化方法至关重要.
    方法:在这项工作中,我们提出了一种新颖的非配对OCT图像去噪方法,称为噪声模拟学习(NIL)。NIL包括三个主要模块:噪声提取模块,通过对噪声图像进行去噪提取噪声特征;噪声模仿模块,合成嘈杂的图像并生成虚假的干净图像;以及对抗性学习模块,通过对抗性训练区分真实和假的干净图像。NIL的复杂度明显低于以前的未配对方法,只利用一个发生器和一个鉴别器进行训练。
    结果:通过有效地融合不成对的图像并采用对抗训练,NIL可以提取更多的斑点噪声信息,以增强去噪性能。建立在NIL上,我们提出了一种OCT图像去噪管道,NIL-NAFNet。这条管道实现了PSNR,SSIM,和RMSE值31.27dB,分别为0.865和7.00,在PKU37数据集上。大量实验表明,我们的方法在定性和定量上都优于最先进的不成对方法。
    结论:这些发现表明,所提出的NIL是一种简单而有效的方法,用于减少不成对的OCT斑点噪声。基于NIL的OCT去噪管道展示了卓越的性能和效率。通过解决斑点噪声,而不需要良好配准的图像对,该方法可以提高图像质量和临床诊断的准确性。
    OBJECTIVE: Optical coherence tomography (OCT) is widely used in clinical practice for its non-invasive, high-resolution imaging capabilities. However, speckle noise inherent to its low coherence principle can degrade image quality and compromise diagnostic accuracy. While deep learning methods have shown promise in reducing speckle noise, obtaining well-registered image pairs remains challenging, leading to the development of unpaired methods. Despite their potential, existing unpaired methods suffer from redundancy in network structures or interaction mechanisms. Therefore, a more streamlined method for unpaired OCT denoising is essential.
    METHODS: In this work, we propose a novel unpaired method for OCT image denoising, referred to as noise-imitation learning (NIL). NIL comprises three primary modules: the noise extraction module, which extracts noise features by denoising noisy images; the noise imitation module, which synthesizes noisy images and generates fake clean images; and the adversarial learning module, which differentiates between real and fake clean images through adversarial training. The complexity of NIL is significantly lower than that of previous unpaired methods, utilizing only one generator and one discriminator for training.
    RESULTS: By efficiently fusing unpaired images and employing adversarial training, NIL can extract more speckle noise information to enhance denoising performance. Building on NIL, we propose an OCT image denoising pipeline, NIL-NAFNet. This pipeline achieved PSNR, SSIM, and RMSE values of 31.27 dB, 0.865, and 7.00, respectively, on the PKU37 dataset. Extensive experiments suggest that our method outperforms state-of-the-art unpaired methods both qualitatively and quantitatively.
    CONCLUSIONS: These findings indicate that the proposed NIL is a simple yet effective method for unpaired OCT speckle noise reduction. The OCT denoising pipeline based on NIL demonstrates exceptional performance and efficiency. By addressing speckle noise without requiring well-registered image pairs, this method can enhance image quality and diagnostic accuracy in clinical practice.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    磁共振成像(MRI)通常用于研究婴儿的大脑发育。然而,由于图像采集时间长,受试者依从性有限,高质量的婴儿MRI可能具有挑战性。在不给图像采集带来额外负担的情况下,图像超分辨率(SR)可用于增强采集后的图像质量。大多数SR技术在多个对齐的低分辨率(LR)和高分辨率(HR)图像对上进行监督和训练,这在实践中通常是不可用的。与监督方法不同,深度图像先验(DIP)可以用于无监督的单图像SR,仅利用输入LR图像进行从头优化以产生HR图像。然而,确定何时在DIP训练早期停止是不平凡的,并且提出了完全自动化SR过程的挑战。为了解决这个问题,我们将SR图像的低频k空间限制为与LR图像相似。我们通过设计一个双模态框架来进一步提高性能,该框架利用T1加权和T2加权图像之间的共享解剖信息。我们评估了我们的模型,双模态DIP(dmDIP),从出生到一岁的婴儿MRI数据,这表明,增强的图像质量可以获得显著降低的敏感性提前停止。
    Magnetic resonance imaging (MRI) is commonly used for studying infant brain development. However, due to the lengthy image acquisition time and limited subject compliance, high-quality infant MRI can be challenging. Without imposing additional burden on image acquisition, image super-resolution (SR) can be used to enhance image quality post-acquisition. Most SR techniques are supervised and trained on multiple aligned low-resolution (LR) and high-resolution (HR) image pairs, which in practice are not usually available. Unlike supervised approaches, Deep Image Prior (DIP) can be employed for unsupervised single-image SR, utilizing solely the input LR image for de novo optimization to produce an HR image. However, determining when to stop early in DIP training is non-trivial and presents a challenge to fully automating the SR process. To address this issue, we constrain the low-frequency k-space of the SR image to be similar to that of the LR image. We further improve performance by designing a dual-modal framework that leverages shared anatomical information between T1-weighted and T2-weighted images. We evaluated our model, dual-modal DIP (dmDIP), on infant MRI data acquired from birth to one year of age, demonstrating that enhanced image quality can be obtained with substantially reduced sensitivity to early stopping.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    针对生命树的先前未探索的部分的基因组测序的最近加速提出了计算挑战。从野外收集的样本通常包含来自几种生物的序列,包括目标,它的杂物,和污染物。因此需要有效的方法来分离序列。尽管测序技术的进步使这项任务变得更容易,仍然难以从数据库中没有很好代表的真核分类单元中分类分配序列。因此,仅基于参考的方法是不够的。这里,我研究了我们如何利用生物体之间序列组成的差异来识别共生体,样本中的寄生虫和污染物,对参考数据的依赖最小。为此,我探索达尔文生命之树项目的数据,包括数百套高质量的HiFi阅读昆虫。可视化由变分自动编码器学习的读段四核苷酸组成的二维表示可以揭示样品的不同组分。用附加信息注释嵌入,比如编码密度,估计覆盖率,或分类标签允许快速评估数据集的内容。这种方法可以扩展到数百万个序列,使探索未组装的阅读集成为可能,即使是大基因组。结合交互式可视化工具,它允许通过基于参考的筛查报告的大部分cobionts被识别。至关重要的是,它还有助于检索缺少合适参考数据的基因组。
    The recent acceleration in genome sequencing targeting previously unexplored parts of the tree of life presents computational challenges. Samples collected from the wild often contain sequences from several organisms, including the target, its cobionts, and contaminants. Effective methods are therefore needed to separate sequences. Though advances in sequencing technology make this task easier, it remains difficult to taxonomically assign sequences from eukaryotic taxa that are not well-represented in databases. Therefore, reference-based methods alone are insufficient. Here, I examine how we can take advantage of differences in sequence composition between organisms to identify symbionts, parasites and contaminants in samples, with minimal reliance on reference data. To this end, I explore data from the Darwin Tree of Life project, including hundreds of high-quality HiFi read sets from insects. Visualising two-dimensional representations of read tetranucleotide composition learned by a Variational Autoencoder can reveal distinct components of a sample. Annotating the embeddings with additional information, such as coding density, estimated coverage, or taxonomic labels allows rapid assessment of the contents of a dataset. The approach scales to millions of sequences, making it possible to explore unassembled read sets, even for large genomes. Combined with interactive visualisation tools, it allows a large fraction of cobionts reported by reference-based screening to be identified. Crucially, it also facilitates retrieving genomes for which suitable reference data are absent.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    近年来,可穿戴传感器和生物电子学的机器学习技术取得了巨大的进步,它在实时传感数据分析中起着至关重要的作用,为个性化医疗提供临床级信息。为此,监督学习和无监督学习算法已经成为强大的工具,允许检测复杂的模式和关系,高维数据集。在这篇评论中,我们的目标是描述可穿戴传感器机器学习的最新进展,专注于算法技术的关键发展,应用程序,以及这种不断发展的景观所固有的挑战。此外,我们强调了机器学习方法提高准确性的潜力,可靠性,和可穿戴传感器数据的可解释性,并讨论这一新兴领域的机会和局限性。最终,我们的工作旨在为这个令人兴奋和快速发展的领域的未来研究工作提供路线图。
    Recent years have witnessed tremendous advances in machine learning techniques for wearable sensors and bioelectronics, which play an essential role in real-time sensing data analysis to provide clinical-grade information for personalized healthcare. To this end, supervised learning and unsupervised learning algorithms have emerged as powerful tools, allowing for the detection of complex patterns and relationships in large, high-dimensional data sets. In this Review, we aim to delineate the latest advancements in machine learning for wearable sensors, focusing on key developments in algorithmic techniques, applications, and the challenges intrinsic to this evolving landscape. Additionally, we highlight the potential of machine-learning approaches to enhance the accuracy, reliability, and interpretability of wearable sensor data and discuss the opportunities and limitations of this emerging field. Ultimately, our work aims to provide a roadmap for future research endeavors in this exciting and rapidly evolving area.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    对比学习已成为无监督表示学习的基石。它的主要范式涉及利用InfoNCE损失的实例区分任务,其中损失已被证明是一种互信息形式。因此,使用互信息作为度量来分析对比学习已成为一种普遍的做法。然而,这种分析方法提出了困难,由于估计互信息的现实世界的应用的必要性。这在其数学基础的优雅和估计的复杂性之间产生了差距,从而阻碍了从互信息分析中获得坚实和有意义的见解的能力。在这项研究中,我们介绍了三种新颖的方法和一些相关的定理,旨在增强互信息分析的严谨性。尽管简单,这些方法可以带来巨大的效用。利用这些方法,我们重新评估了对比学习分析的三个实例,说明所提出的方法的能力,以促进更深入的理解或纠正先前存在的误解。主要研究结果如下:(1)小批量影响训练损失的范围,它们不会固有地限制学习表示的信息内容或对下游性能产生不利影响;(2)互信息,仔细选择积极的配对和训练后的估计,被证明是评估实际网络的优越措施;(3)区分与任务相关的信息和不相关的信息提出了挑战,然而,不相关的信息源不一定会损害下游任务的概括。
    Contrastive learning has emerged as a cornerstone in unsupervised representation learning. Its primary paradigm involves an instance discrimination task utilizing InfoNCE loss where the loss has been proven to be a form of mutual information. Consequently, it has become a common practice to analyze contrastive learning using mutual information as a measure. Yet, this analysis approach presents difficulties due to the necessity of estimating mutual information for real-world applications. This creates a gap between the elegance of its mathematical foundation and the complexity of its estimation, thereby hampering the ability to derive solid and meaningful insights from mutual information analysis. In this study, we introduce three novel methods and a few related theorems, aimed at enhancing the rigor of mutual information analysis. Despite their simplicity, these methods can carry substantial utility. Leveraging these approaches, we reassess three instances of contrastive learning analysis, illustrating the capacity of the proposed methods to facilitate deeper comprehension or to rectify pre-existing misconceptions. The main results can be summarized as follows: (1) While small batch sizes influence the range of training loss, they do not inherently limit learned representation\'s information content or affect downstream performance adversely; (2) Mutual information, with careful selection of positive pairings and post-training estimation, proves to be a superior measure for evaluating practical networks; and (3) Distinguishing between task-relevant and irrelevant information presents challenges, yet irrelevant information sources do not necessarily compromise the generalization of downstream tasks.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    无线传感器网络(WSN)对于广泛的应用是必不可少的,包括环境监测和智慧城市发展,由于他们能够收集和传输各种物理和环境数据。WSN的性质,再加上具有成本效益的传感器的可变性和噪声敏感性,在实现准确的数据分析和异常检测方面提出了重大挑战。为了解决这些问题,本文提出了一个新的框架,称为在线自适应卡尔曼滤波(OAKF),专为WSN中的实时异常检测而设计。该框架通过响应实时数据动态调整过滤参数和异常检测阈值而脱颖而出,确保在传感器噪声和环境变化中进行准确可靠的异常识别。通过突出计算效率和可扩展性,OAKF框架已针对资源受限的传感器节点进行了优化。对不同WSN数据集大小的验证证实了其有效性,在减少假阳性和阴性以及实现每个样品0.008s的处理时间方面显示95.4%的准确度。
    Wireless sensor networks (WSNs) are essential for a wide range of applications, including environmental monitoring and smart city developments, thanks to their ability to collect and transmit diverse physical and environmental data. The nature of WSNs, coupled with the variability and noise sensitivity of cost-effective sensors, presents significant challenges in achieving accurate data analysis and anomaly detection. To address these issues, this paper presents a new framework, called Online Adaptive Kalman Filtering (OAKF), specifically designed for real-time anomaly detection within WSNs. This framework stands out by dynamically adjusting the filtering parameters and anomaly detection threshold in response to live data, ensuring accurate and reliable anomaly identification amidst sensor noise and environmental changes. By highlighting computational efficiency and scalability, the OAKF framework is optimized for use in resource-constrained sensor nodes. Validation on different WSN dataset sizes confirmed its effectiveness, showing 95.4% accuracy in reducing false positives and negatives as well as achieving a processing time of 0.008 s per sample.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    自监督对比学习借鉴权力表示模型,从无标记数据中获取通用语义特征,训练这种模型的关键在于跟踪运动特征的准确性。以前的视频对比学习方法已经广泛使用空间或时间上的增强作为类似的实例,导致模型更有可能学习静态背景而不是运动特征。为了缓解背景快捷方式,在本文中,我们提出了一个跨视图运动一致(CVMC)自监督视频内部对比模型,专注于学习局部细节和长期时间关系。具体来说,我们首先提取连续视频片段的动态特征,然后基于多视图运动一致性对齐这些特征。同时,我们比较了优化的动态特征,用于实例比较不同视频和同一视频中具有时间顺序的局部空间细粒度,分别。最终,时空对齐和运动判别的联合优化有效填补了实例识别缺失组件的挑战,空间紧密度,自我监督学习中的时间感知。实验结果表明,与其他最先进的方法相比,我们提出的自监督模型可以有效地学习视觉表示信息,并在动作识别和视频检索任务中获得高度竞争力。
    Self-supervised contrastive learning draws on power representational models to acquire generic semantic features from unlabeled data, and the key to training such models lies in how accurately to track motion features. Previous video contrastive learning methods have extensively used spatially or temporally augmentation as similar instances, resulting in models that are more likely to learn static backgrounds than motion features. To alleviate the background shortcuts, in this paper, we propose a cross-view motion consistent (CVMC) self-supervised video inter-intra contrastive model to focus on the learning of local details and long-term temporal relationships. Specifically, we first extract the dynamic features of consecutive video snippets and then align these features based on multi-view motion consistency. Meanwhile, we compare the optimized dynamic features for instance comparison of different videos and local spatial fine-grained with temporal order in the same video, respectively. Ultimately, the joint optimization of spatio-temporal alignment and motion discrimination effectively fills the challenges of the missing components of instance recognition, spatial compactness, and temporal perception in self-supervised learning. Experimental results show that our proposed self-supervised model can effectively learn visual representation information and achieve highly competitive performance compared to other state-of-the-art methods in both action recognition and video retrieval tasks.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    无监督语义分割对于理解每个像素属于没有注释的已知类别是重要的。最近的研究已经证明了有希望的结果,通过采用视觉变压器骨干预训练在图像级数据集以自我监督的方式。然而,这些方法总是依赖于复杂的架构或精心设计的输入。自然,我们正试图用一种简单的方法来探索投资。为了防止过度并发症,本文介绍了一种简单的密集嵌入对比网络(DECNet),用于无监督语义分割。具体来说,我们提出了一种最近邻相似性策略(NNS),以建立定义明确的正负对进行密集对比学习。同时,我们优化了一个名为Ortho-InfoNCE的对比目标,以缓解对比学习中固有的假阴性问题,从而进一步增强密集表示。最后,在COCO-Stuff和Cityscapes数据集上进行的大量实验表明,我们的方法优于最先进的方法。
    Unsupervised semantic segmentation is important for understanding that each pixel belongs to known categories without annotation. Recent studies have demonstrated promising outcomes by employing a vision transformer backbone pre-trained on an image-level dataset in a self-supervised manner. However, those methods always depend on complex architectures or meticulously designed inputs. Naturally, we are attempting to explore the investment with a straightforward approach. To prevent over-complication, we introduce a simple Dense Embedding Contrast network (DECNet) for unsupervised semantic segmentation in this paper. Specifically, we propose a Nearest Neighbor Similarity strategy (NNS) to establish well-defined positive and negative pairs for dense contrastive learning. Meanwhile, we optimize a contrastive objective named Ortho-InfoNCE to alleviate the false negative problem inherent in contrastive learning for further enhancing dense representations. Finally, extensive experiments conducted on COCO-Stuff and Cityscapes datasets demonstrate that our approach outperforms state-of-the-art methods.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

公众号