zero-shot

零射
  • 文章类型: Journal Article
    分割任意模型(SAM)是最近开发的用于图像分割的全范围基础模型。它可以使用稀疏的手动提示(如边界框)在自然图像中生成像素级分割,但在医学图像(如低对比度,嘈杂的超声图像。我们提出了一种改进的测试阶段提示增强技术,旨在提高SAM在医学图像分割中的性能。该方法将多框提示增强与基于aleatoric不确定性的假阴性(FN)和假阳性(FP)校正(FNPC)策略相结合。我们在两个超声数据集上评估了该方法,并显示了SAM的性能和对不准确提示的鲁棒性的改进,无需进一步培训或调整。此外,我们提出了单片体积(SS2V)方法,仅使用来自单个2D切片的边界框注释来启用3D像素级分割。我们的结果允许在甚至嘈杂的情况下有效使用SAM,低对比度医学图像。源代码已发布于:https://github.com/MedICL-VU/FNPC-SAM。
    The Segment Anything Model (SAM) is a recently developed all-range foundation model for image segmentation. It can use sparse manual prompts such as bounding boxes to generate pixel-level segmentation in natural images but struggles in medical images such as low-contrast, noisy ultrasound images. We propose a refined test-phase prompt augmentation technique designed to improve SAM\'s performance in medical image segmentation. The method couples multi-box prompt augmentation and an aleatoric uncertainty-based false-negative (FN) and false-positive (FP) correction (FNPC) strategy. We evaluate the method on two ultrasound datasets and show improvement in SAM\'s performance and robustness to inaccurate prompts, without the necessity for further training or tuning. Moreover, we present the Single-Slice-to-Volume (SS2V) method, enabling 3D pixel-level segmentation using only the bounding box annotation from a single 2D slice. Our results allow efficient use of SAM in even noisy, low-contrast medical images. The source code has been released at: https://github.com/MedICL-VU/FNPC-SAM.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    背景:大型语言模型(LLM)在自然语言处理(NLP)中显示出非凡的能力,特别是在标记数据稀缺或昂贵的领域,例如临床领域。然而,为了解开隐藏在这些LLM中的临床知识,我们需要设计有效的提示,引导他们在没有任何任务特定训练数据的情况下执行特定的临床NLP任务.这被称为上下文学习,这是一门艺术和科学,需要了解不同LLM的优势和劣势,并迅速采用工程方法。
    目的:本研究的目的是评估各种即时工程技术的有效性,包括2个新引入的类型-启发式和合奏提示,使用预训练的语言模型进行零射和少射临床信息提取。
    方法:这项全面的实验研究评估了不同的提示类型(简单的前缀,简单的完形填空,思想链,预期,启发式,和合奏)跨越5个临床NLP任务:临床意义消歧,生物医学证据提取,共同参照决议,药物状态提取,和药物属性提取。使用3种最先进的语言模型评估了这些提示的性能:GPT-3.5(OpenAI),双子座(谷歌),和LLaMA-2(Meta)。该研究将零射与少射提示进行了对比,并探讨了合奏方法的有效性。
    结果:研究表明,针对特定任务的提示定制对于LLM在零射临床NLP中的高性能至关重要。在临床意义上的消歧,GPT-3.5在启发式提示下达到0.96的准确性,在生物医学证据提取中达到0.94的准确性。启发式提示,伴随着一连串的思想提示,跨任务非常有效。在复杂的场景中,很少有机会提示提高性能,和集合方法利用了多种即时优势。GPT-3.5在任务和提示类型上的表现始终优于Gemini和LLaMA-2。
    结论:本研究对即时工程方法进行了严格的评估,并介绍了临床信息提取的创新技术,证明了临床领域上下文学习的潜力。这些发现为未来基于提示的临床NLP研究提供了明确的指导方针。促进非NLP专家参与临床NLP进步。据我们所知,这是在这个生成人工智能时代,对临床NLP的不同提示工程方法进行实证评估的首批作品之一,我们希望它能激励和指导未来在这一领域的研究。
    BACKGROUND: Large language models (LLMs) have shown remarkable capabilities in natural language processing (NLP), especially in domains where labeled data are scarce or expensive, such as the clinical domain. However, to unlock the clinical knowledge hidden in these LLMs, we need to design effective prompts that can guide them to perform specific clinical NLP tasks without any task-specific training data. This is known as in-context learning, which is an art and science that requires understanding the strengths and weaknesses of different LLMs and prompt engineering approaches.
    OBJECTIVE: The objective of this study is to assess the effectiveness of various prompt engineering techniques, including 2 newly introduced types-heuristic and ensemble prompts, for zero-shot and few-shot clinical information extraction using pretrained language models.
    METHODS: This comprehensive experimental study evaluated different prompt types (simple prefix, simple cloze, chain of thought, anticipatory, heuristic, and ensemble) across 5 clinical NLP tasks: clinical sense disambiguation, biomedical evidence extraction, coreference resolution, medication status extraction, and medication attribute extraction. The performance of these prompts was assessed using 3 state-of-the-art language models: GPT-3.5 (OpenAI), Gemini (Google), and LLaMA-2 (Meta). The study contrasted zero-shot with few-shot prompting and explored the effectiveness of ensemble approaches.
    RESULTS: The study revealed that task-specific prompt tailoring is vital for the high performance of LLMs for zero-shot clinical NLP. In clinical sense disambiguation, GPT-3.5 achieved an accuracy of 0.96 with heuristic prompts and 0.94 in biomedical evidence extraction. Heuristic prompts, alongside chain of thought prompts, were highly effective across tasks. Few-shot prompting improved performance in complex scenarios, and ensemble approaches capitalized on multiple prompt strengths. GPT-3.5 consistently outperformed Gemini and LLaMA-2 across tasks and prompt types.
    CONCLUSIONS: This study provides a rigorous evaluation of prompt engineering methodologies and introduces innovative techniques for clinical information extraction, demonstrating the potential of in-context learning in the clinical domain. These findings offer clear guidelines for future prompt-based clinical NLP research, facilitating engagement by non-NLP experts in clinical NLP advancements. To the best of our knowledge, this is one of the first works on the empirical evaluation of different prompt engineering approaches for clinical NLP in this era of generative artificial intelligence, and we hope that it will inspire and inform future research in this area.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    在巨大的组合突变环境中以最少的实验努力获得最大数量的表现最佳的变体的策略将在提高蛋白质工程的资源生产能力方面具有巨大的效用。为了这个目标,我们提出了一种简单有效的基于机器学习的策略,优于其他最先进的方法。我们的策略集成了零拍预测和多轮采样,通过仅尝试一些预测的顶级变体来指导主动学习。我们发现,机器学习的12个变体的四轮低N挑选和验证采样在选择组合突变文库中真正的前1%变体时产生了高达92.6%的最佳准确度。而24个变体的两轮也可以使用。我们展示了我们的策略,成功地发现了来自不同家族的高性能蛋白质变体,包括基于CRISPR的基因组编辑器。支持其用于解决蛋白质工程任务的可推广应用。补充信息中包含了本文透明的同行评审过程的记录。
    A strategy to obtain the greatest number of best-performing variants with least amount of experimental effort over the vast combinatorial mutational landscape would have enormous utility in boosting resource producibility for protein engineering. Toward this goal, we present a simple and effective machine learning-based strategy that outperforms other state-of-the-art methods. Our strategy integrates zero-shot prediction and multi-round sampling to direct active learning via experimenting with only a few predicted top variants. We find that four rounds of low-N pick-and-validate sampling of 12 variants for machine learning yielded the best accuracy of up to 92.6% in selecting the true top 1% variants in combinatorial mutant libraries, whereas two rounds of 24 variants can also be used. We demonstrate our strategy in successfully discovering high-performance protein variants from diverse families including the CRISPR-based genome editors, supporting its generalizable application for solving protein engineering tasks. A record of this paper\'s transparent peer review process is included in the supplemental information.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    随着deepfake技术的发展,deepfake检测受到了广泛的关注。尽管已经提出了一些deepfake取证技术,它们仍然很难在现实世界的场景中实现。这是由于不同的deepfake技术以及传播过程中视频的压缩或编辑的差异。考虑到Deepfake检测中样本不平衡与少量拍摄场景的问题,提出了一种基于元学习的多特征信道域加权框架(MCW)。为了获得卓越的跨数据库检测性能,提出的框架从两个方面改进了元学习网络:它通过结合图像的RGB域和频域信息来增强模型检测目标的特征提取能力,并通过向特征映射上的通道分配元权重来增强模型检测目标的泛化能力。提出的MCW框架解决了算法对未知算法生成的样本的检测性能差和抗数据压缩能力不足的问题。实验是在零镜头和少镜头的情况下进行的,在真实情况下模拟deepfake检测环境。我们选择了9种检测算法作为比较算法。实验结果表明,MCW框架在跨算法检测和跨数据集检测方面优于其他算法。MCW框架展示了其利用低质量训练图像和跨不同生成算法场景来概括和抵抗压缩的能力,它在少量学习场景中具有更好的微调潜力。
    With the development of deepfake technology, deepfake detection has received widespread attention. Although some deepfake forensics techniques have been proposed, they are still very difficult to implement in real-world scenarios. This is due to the differences in different deepfake technologies and the compression or editing of videos during the propagation process. Considering the issue of sample imbalance with few-shot scenarios in deepfake detection, we propose a multi-feature channel domain-weighted framework based on meta-learning (MCW). In order to obtain outstanding detection performance of a cross-database, the proposed framework improves a meta-learning network in two ways: it enhances the model\'s feature extraction ability for detecting targets by combining the RGB domain and frequency domain information of the image and enhances the model\'s generalization ability for detecting targets by assigning meta weights to channels on the feature map. The proposed MCW framework solves the problems of poor detection performance and insufficient data compression resistance of the algorithm for samples generated by unknown algorithms. The experiment was set in a zero-shot scenario and few-shot scenario, simulating the deepfake detection environment in real situations. We selected nine detection algorithms as comparative algorithms. The experimental results show that the MCW framework outperforms other algorithms in cross-algorithm detection and cross-dataset detection. The MCW framework demonstrates its ability to generalize and resist compression with low-quality training images and across different generation algorithm scenarios, and it has better fine-tuning potential in few-shot learning scenarios.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    在恶劣环境照明下拍摄的照片可能由于曝光不足而遭受许多图像质量劣化现象。这些包括降低亮度,传输信息丢失,噪音,和颜色失真。为了解决上述问题,研究人员提出了许多基于深度学习的方法来提高图像的照度。然而,现有的大多数方法都面临着难以获取成对训练数据的问题。在这种情况下,本文提出了一种低光照条件下的零参考图像增强网络。首先,改进的编码器-解码器结构用于提取图像特征以生成特征图,并从特征图生成增强因子的参数矩阵。然后,使用参数矩阵构建增强曲线。使用增强曲线和增强参数迭代地增强图像。第二,无监督算法在训练中需要设计图像非参考损失函数。引入四个非参考损失函数来训练参数估计网络。在几个只有弱光图像的数据集上的实验表明,与NIQE中的其他方法相比,该网络的性能有所提高。PIQE,和BRISQUE非参考评价指标,并对关键部位进行了烧蚀实验,证明了该方法的有效性。同时,研究了该方法在PC设备和移动设备上的性能数据,并进行了实验分析。这证明了本文方法在实际应用中的可行性。
    Photographs taken under harsh ambient lighting can suffer from a number of image quality degradation phenomena due to insufficient exposure. These include reduced brightness, loss of transfer information, noise, and color distortion. In order to solve the above problems, researchers have proposed many deep learning-based methods to improve the illumination of images. However, most existing methods face the problem of difficulty in obtaining paired training data. In this context, a zero-reference image enhancement network for low light conditions is proposed in this paper. First, the improved Encoder-Decoder structure is used to extract image features to generate feature maps and generate the parameter matrix of the enhancement factor from the feature maps. Then, the enhancement curve is constructed using the parameter matrix. The image is iteratively enhanced using the enhancement curve and the enhancement parameters. Second, the unsupervised algorithm needs to design an image non-reference loss function in training. Four non-reference loss functions are introduced to train the parameter estimation network. Experiments on several datasets with only low-light images show that the proposed network has improved performance compared with other methods in NIQE, PIQE, and BRISQUE non-reference evaluation index, and ablation experiments are carried out for key parts, which proves the effectiveness of this method. At the same time, the performance data of the method on PC devices and mobile devices are investigated, and the experimental analysis is given. This proves the feasibility of the method in this paper in practical application.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    零镜头域适应(ZDA)方法旨在将有关在源域中学习的任务的知识转移到目标域,而来自目标域的任务相关数据不可用。在这项工作中,考虑到ZDA的任务特征,我们解决了学习特征表示,这些特征表示对于不同领域是不变的,并且在不同领域之间共享。为此,我们提出了一种任务引导的ZDA(TG-ZDA)方法,该方法采用多分支深度神经网络来学习特征表示,利用其域不变性和可共享性。所提出的TG-ZDA模型可以被端到端地训练,而不需要从目标域的估计表示生成的合成任务和数据。已在图像分类数据集上使用基准ZDA任务检查了所提出的TG-ZDA。实验结果表明,对于不同的领域和任务,我们提出的TG-ZDA优于最先进的ZDA方法。
    Zero-shot domain adaptation (ZDA) methods aim to transfer knowledge about a task learned in a source domain to a target domain, while task-relevant data from target domain are not available. In this work, we address learning feature representations which are invariant to and shared among different domains considering task characteristics for ZDA. To this end, we propose a method for task-guided ZDA (TG-ZDA) which employs multi-branch deep neural networks to learn feature representations exploiting their domain invariance and shareability properties. The proposed TG-ZDA models can be trained end-to-end without requiring synthetic tasks and data generated from estimated representations of target domains. The proposed TG-ZDA has been examined using benchmark ZDA tasks on image classification datasets. Experimental results show that our proposed TG-ZDA outperforms state-of-the-art ZDA methods for different domains and tasks.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    以前在图像去雾中使用的大多数基于学习的方法都采用监督学习策略,这是耗时的,需要一个大规模的数据集。然而,大规模数据集是很难获得的。这里,我们提出了一种基于暗通道先验的自监督零发去雾网络(SZDNet),它使用从输出的去雾图像生成的模糊图像作为伪标签来监督网络的优化过程。此外,我们使用一种新颖的多通道四叉树算法来估计大气光值,比以前的方法更准确。此外,伪标签和输入图像之间的余弦距离和均方误差之和作为损失函数来提高去雾图像的质量。SZDNet最显著的优点是,在执行去雾任务之前,它不需要大型数据集进行训练。与最先进的方法相比,广泛的测试表明,所提出的方法在定性和定量评估中都具有良好的性能。
    Most learning-based methods previously used in image dehazing employ a supervised learning strategy, which is time-consuming and requires a large-scale dataset. However, large-scale datasets are difficult to obtain. Here, we propose a self-supervised zero-shot dehazing network (SZDNet) based on dark channel prior, which uses a hazy image generated from the output dehazed image as a pseudo-label to supervise the optimization process of the network. Additionally, we use a novel multichannel quad-tree algorithm to estimate atmospheric light values, which is more accurate than previous methods. Furthermore, the sum of the cosine distance and the mean squared error between the pseudo-label and the input image is applied as a loss function to enhance the quality of the dehazed image. The most significant advantage of the SZDNet is that it does not require a large dataset for training before performing the dehazing task. Extensive testing shows promising performances of the proposed method in both qualitative and quantitative evaluations when compared with state-of-the-art methods.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    立场检测的主要挑战是大量(可能是无限的)和多样化的立场主题。由于注释的费用和新的现实世界主题的不断创建,为这样的集合收集数据是不现实的(例如,一位新政客竞选公职)。此外,稳定发生在广泛的语言和流派中(例如,Twitter,新闻文章)。虽然英语中的零射姿态检测,评估是在培训期间没有看到的主题,受到越来越多的关注,我们认为,这种关注应该扩大到多语言和多流派的设置。我们讨论了英语零射姿态检测评估的两种范式,以及最近在这方面的工作。然后,我们讨论了多语言和多流派立场检测的最新工作,主要关注非零镜头设置。我们认为,这项工作应该扩展到多语言和多类型的零射姿态检测,并提出最佳实践,以系统化和刺激未来的工作。虽然域自适应技术非常适合在这些设置中工作,我们认为,应该更加注意提高模型的可解释性,并进行强有力的评估,不仅要考虑经验泛化能力,还要考虑对复杂语言和推论的理解。
    A major challenge in stance detection is the large (potentially infinite) and diverse set of stance topics. Collecting data for such a set is unrealistic due to both the expense of annotation and the continuous creation of new real-world topics (e.g., a new politician runs for office). Furthermore, stancetaking occurs in a wide range of languages and genres (e.g., Twitter, news articles). While zero-shot stance detection in English, where evaluation is on topics not seen during training, has received increasing attention, we argue that this attention should be expanded to multilingual and multi-genre settings. We discuss two paradigms for English zero-shot stance detection evaluation, as well as recent work in this area. We then discuss recent work on multilingual and multi-genre stance detection, which has focused primarily on non-zero-shot settings. We argue that this work should be expanded to multilingual and multi-genre zero-shot stance detection and propose best practices to systematize and stimulate future work in this direction. While domain adaptation techniques are well-suited for work in these settings, we argue that increased care should be taken to improve model explainability and to conduct robust evaluations, considering not only empirical generalization ability but also the understanding of complex language and inferences.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    智力的一个重要方面是在没有任何直接经验的情况下适应一项新任务的能力(零射击),基于它与以前任务的关系。人类可以表现出这种认知灵活性。相比之下,在特定任务中实现超人表现的模型通常无法适应甚至轻微的任务变化。为了解决这个问题,我们提出了一个通用的计算框架,用于根据新任务与先前任务的关系来适应新任务。我们首先学习任务的矢量表示。适应新的任务,我们提出了元映射,转换基本任务表示的高阶任务。我们证明了这个框架在各种任务和计算范式中的有效性,从回归到图像分类和强化学习。我们比较了人类的适应性和基于语言的零镜头学习方法。在这些领域,元映射是成功的,通常达到80%至90%的性能,没有任何数据,在一个新奇的任务中,即使新任务与先前的经验直接矛盾。我们进一步证明,元映射不仅可以通过学习的关系推广到新任务,但也可以使用在训练中看不见的新关系来概括。最后,使用元映射作为起点可以大大加快以后对新任务的学习,并大大减少学习时间和累积误差。我们的结果为智能适应性的可能计算基础提供了见解,并为建模认知灵活性和构建更灵活的人工智能系统提供了可能的框架。
    An important aspect of intelligence is the ability to adapt to a novel task without any direct experience (zero shot), based on its relationship to previous tasks. Humans can exhibit this cognitive flexibility. By contrast, models that achieve superhuman performance in specific tasks often fail to adapt to even slight task alterations. To address this, we propose a general computational framework for adapting to novel tasks based on their relationship to prior tasks. We begin by learning vector representations of tasks. To adapt to new tasks, we propose metamappings, higher-order tasks that transform basic task representations. We demonstrate the effectiveness of this framework across a wide variety of tasks and computational paradigms, ranging from regression to image classification and reinforcement learning. We compare to both human adaptability and language-based approaches to zero-shot learning. Across these domains, metamapping is successful, often achieving 80 to 90% performance, without any data, on a novel task, even when the new task directly contradicts prior experience. We further show that metamapping can not only generalize to new tasks via learned relationships, but can also generalize using novel relationships unseen during training. Finally, using metamapping as a starting point can dramatically accelerate later learning on a new task and reduce learning time and cumulative error substantially. Our results provide insight into a possible computational basis of intelligent adaptability and offer a possible framework for modeling cognitive flexibility and building more flexible artificial intelligence systems.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Sci-hub)

       PDF(Pubmed)

公众号