Deep metric learning

  • 文章类型: Journal Article
    未知疾病的出现通常很少或没有可用的样品。零射学习和少射学习在医学图像分析中具有广阔的应用前景。在本文中,我们提出了一种跨模态深度度量学习广义零分学习(CM-DML-GZSL)模型。拟议的网络由视觉特征提取器组成,固定的语义特征提取器,和深度回归模块。该网络属于用于多种模态的双流网络。在多标签设置中,每个样本平均包含少量阳性标签和大量阴性标签。这种正负不平衡主导了优化过程,并可能阻止在训练期间建立视觉特征和语义向量之间的有效对应关系。导致精度较低。在这方面引入了一种新颖的加权聚焦欧几里得距离度量损失。这种损失不仅可以动态增加硬质样品的重量,而且可以减少简单样品的重量,但它也可以促进样本和与其正标签相对应的语义向量之间的联系,这有助于减轻在广义零拍学习设置中预测看不见的类的偏差。加权聚焦欧氏距离度量损失函数可以动态调整样本权重,为胸部X光诊断提供零拍多标签学习,正如在大型公开数据集上的实验结果表明的那样。
    The emergence of unknown diseases is often with few or no samples available. Zero-shot learning and few-shot learning have promising applications in medical image analysis. In this paper, we propose a Cross-Modal Deep Metric Learning Generalized Zero-Shot Learning (CM-DML-GZSL) model. The proposed network consists of a visual feature extractor, a fixed semantic feature extractor, and a deep regression module. The network belongs to a two-stream network for multiple modalities. In a multi-label setting, each sample contains a small number of positive labels and a large number of negative labels on average. This positive-negative imbalance dominates the optimization procedure and may prevent the establishment of an effective correspondence between visual features and semantic vectors during training, resulting in a low degree of accuracy. A novel weighted focused Euclidean distance metric loss is introduced in this regard. This loss not only can dynamically increase the weight of hard samples and decrease the weight of simple samples, but it can also promote the connection between samples and semantic vectors corresponding to their positive labels, which helps mitigate bias in predicting unseen classes in the generalized zero-shot learning setting. The weighted focused Euclidean distance metric loss function can dynamically adjust sample weights, enabling zero-shot multi-label learning for chest X-ray diagnosis, as experimental results on large publicly available datasets demonstrate.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    单细胞RNA测序(scRNA-seq)显着加速了复杂组织和生物体中不同细胞谱系和类型的实验表征。细胞类型注释在大多数scRNA-seq分析管道中非常重要。然而,手动细胞类型注释很大程度上依赖于scRNA-seq数据和标记基因的质量,因此可能是费力和耗时的。此外,scRNA-seq数据集的异质性对准确的细胞类型注释提出了另一个挑战,例如由不同的scRNA-seq方案和样品诱导的分批效应。为了克服这些限制,在这里,我们提出了一个新的管道,叫做TripletCell,对于跨物种,跨方案和跨样本细胞类型注释。我们在TripletCell中开发了用于特征提取(FE)的单元嵌入和降维模块,即TripletCell-FE,利用基于深度度量学习的算法来处理参考基因表达矩阵和查询细胞之间的关系。我们对21个数据集的实验研究(涵盖9个scRNA-seq方案,两个物种和三个组织)证明TripletCell优于最先进的细胞类型注释方法。更重要的是,不管协议或物种,TripletCell可以在注释不同类型的细胞方面提供出色而强大的性能。TripletCell可在https://github.com/liuyan3056/TripletCell上免费获得。我们相信TripletCell是使用scRNA-seq数据准确注释各种细胞类型的可靠计算工具,并将有助于在细胞生物学中产生新的生物学假设。
    Single-cell RNA sequencing (scRNA-seq) has significantly accelerated the experimental characterization of distinct cell lineages and types in complex tissues and organisms. Cell-type annotation is of great importance in most of the scRNA-seq analysis pipelines. However, manual cell-type annotation heavily relies on the quality of scRNA-seq data and marker genes, and therefore can be laborious and time-consuming. Furthermore, the heterogeneity of scRNA-seq datasets poses another challenge for accurate cell-type annotation, such as the batch effect induced by different scRNA-seq protocols and samples. To overcome these limitations, here we propose a novel pipeline, termed TripletCell, for cross-species, cross-protocol and cross-sample cell-type annotation. We developed a cell embedding and dimension-reduction module for the feature extraction (FE) in TripletCell, namely TripletCell-FE, to leverage the deep metric learning-based algorithm for the relationships between the reference gene expression matrix and the query cells. Our experimental studies on 21 datasets (covering nine scRNA-seq protocols, two species and three tissues) demonstrate that TripletCell outperformed state-of-the-art approaches for cell-type annotation. More importantly, regardless of protocols or species, TripletCell can deliver outstanding and robust performance in annotating different types of cells. TripletCell is freely available at https://github.com/liuyan3056/TripletCell. We believe that TripletCell is a reliable computational tool for accurately annotating various cell types using scRNA-seq data and will be instrumental in assisting the generation of novel biological hypotheses in cell biology.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    近年来,已经报道了越来越多的基于深度自动编码器的智能状态监测和异常检测算法,以提高风力涡轮机的可靠性。然而,现有的研究大多只关注于无监督方式下正常数据的精确建模,很少有研究在学习过程中利用故障实例的信息,这导致次优的检测性能和较低的鲁棒性。为此,我们首先开发了一个由故障实例增强的深度自动编码器,也就是说,三元组卷积深度自动编码器(三元组-ConvDAE),联合集成卷积自动编码器和深度度量学习。在故障实例的帮助下,三元组-ConvDAE不仅可以捕获正常的操作数据模式,还可以获得有区别的深度嵌入特征。此外,为了克服稀缺故障实例的困难,我们采用了一种改进的基于生成对抗网络的数据增强方法来生成高质量的合成故障实例。最后,我们使用多种性能度量验证了所提出的异常检测方法的性能。实验结果表明,我们的方法优于其他三种最先进的方法。此外,当故障实例不足时,所提出的增强方法可以有效地提高三元组-ConvDAE的性能。
    An increasing number of deep autoencoder-based algorithms for intelligent condition monitoring and anomaly detection have been reported in recent years to improve wind turbine reliability. However, most existing studies have only focused on the precise modeling of normal data in an unsupervised manner; few studies have utilized the information of fault instances in the learning process, which results in suboptimal detection performance and low robustness. To this end, we first developed a deep autoencoder enhanced by fault instances, that is, a triplet-convolutional deep autoencoder (triplet-Conv DAE), jointly integrating a convolutional autoencoder and deep metric learning. Aided by fault instances, triplet-Conv DAE can not only capture normal operation data patterns but also acquire discriminative deep embedding features. Moreover, to overcome the difficulty of scarce fault instances, we adopted an improved generative adversarial network-based data augmentation method to generate high-quality synthetic fault instances. Finally, we validated the performance of the proposed anomaly detection method using a multitude of performance measures. The experimental results show that our method is superior to three other state-of-the-art methods. In addition, the proposed augmentation method can efficiently improve the performance of the triplet-Conv DAE when fault instances are insufficient.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    -胸部疾病,像许多其他疾病一样,会导致并发症。现有的多标签医学图像学习问题通常包含丰富的病理信息,如图像,属性,和标签,这对于补充临床诊断至关重要。然而,大多数当代的努力都集中在从输入到二进制标签的回归上,忽略标签的视觉特征和语义向量之间的关系。此外,疾病之间的数据量不平衡,这经常导致智能诊断系统做出错误的疾病预测。因此,我们旨在提高胸部X线图像多标签分类的准确性。胸部X-射线14图片被用作本研究中的实验的多标签数据集。通过微调ConvNeXt网络,我们得到了视觉向量,我们将其与BioBert编码的语义向量相结合,将两种不同形式的特征映射到公共度量空间中,并使语义向量成为度量空间中每个类的原型。然后从图像级别和疾病类别级别考虑图像和标签之间的度量关系,分别,提出了一种新的双加权度量损失函数。最后,实验中获得的平均AUC评分达到0.826,我们的模型优于对比模型.
    -Thoracic disease, like many other diseases, can lead to complications. Existing multi-label medical image learning problems typically include rich pathological information, such as images, attributes, and labels, which are crucial for supplementary clinical diagnosis. However, the majority of contemporary efforts exclusively focus on regression from input to binary labels, ignoring the relationship between visual features and semantic vectors of labels. In addition, there is an imbalance in data amount between diseases, which frequently causes intelligent diagnostic systems to make erroneous disease predictions. Therefore, we aim to improve the accuracy of the multi-label classification of chest X-ray images. Chest X-ray14 pictures were utilized as the multi-label dataset for the experiments in this study. By fine-tuning the ConvNeXt network, we got visual vectors, which we combined with semantic vectors encoded by BioBert to map the two different forms of features into a common metric space and made semantic vectors the prototype of each class in metric space. The metric relationship between images and labels is then considered from the image level and disease category level, respectively, and a new dual-weighted metric loss function is proposed. Finally, the average AUC score achieved in the experiment reached 0.826, and our model outperformed the comparison models.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    Melanoma is a tumor caused by melanocytes with a high degree of malignancy, easy local recurrence, distant metastasis, and poor prognosis. It is also difficult to be detected by inexperienced dermatologist due to their similar appearances, such as color, shape, and contour.
    To develop and test a new computer-aided diagnosis scheme to detect melanoma skin cancer.
    In this new scheme, the unsupervised clustering based on deep metric learning is first conducted to make images with high similarity together and the corresponding model weights are utilized as teacher-model for the next stage. Second, benefit from the knowledge distillation, the attention transfer is adopted to make the classification model enable to learn the similarity features and information of categories simultaneously which improve the diagnosis accuracy than the common classification method.
    In validation sets, 8 categories were included, and 2443 samples were calculated. The highest accuracy of the new scheme is 0.7253, which is 5% points higher than the baseline (0.6794). Specifically, the F1-Score of three malignant lesions BCC (Basal cell carcinoma), SCC (Squamous cell carcinomas), and MEL (Melanoma) increase from 0.65 to 0.73, 0.28 to 0.37, and 0.54 to 0.58, respectively. In two test sets of HAN including 3844 samples and BCN including 6375 samples, the highest accuracies are 0.68 and 0.53 for HAM and BCN datasets, respectively, which are higher than the baseline (0.649 and 0.516). Additionally, F1 scores of BCC, SCC, MEL are 0.49, 0.2, 0.45 in HAM dataset and 0.6, 0.14, 0.55 in BCN dataset, respectively, which are also higher than F1 scores the results of baseline.
    This study demonstrates that the similarity clustering method enables to extract the related feature information to gather similar images together. Moreover, based on the attention transfer, the proposed classification framework can improve total accuracy and F1-score of skin lesion diagnosis.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    Objectives.胸部电阻抗断层扫描(EIT)测量中的心脏相关成分对肺灌注监测和心功能测量具有潜在价值。在自发呼吸的情况下,心脏相关信号会受到通气相关信号的严重干扰。传统的心脏相关信号分离方法通常基于信号的某些特征。进一步提高分离精度,应该利用信号的更全面的特征。方法。我们提出了一种无监督的深度学习方法,称为深度特征域匹配(DFDM),其利用期望信号和屏气信号的特征域相似性。该方法的特征在于两个子步骤。第一步,设计并训练了一个新颖的暹罗网络,以学习屏气信号的共同特征;在第二步中,Siamese网络用作分离信号和屏气信号之间的特征匹配约束。主要结果。该方法首先使用合成数据进行测试,结果显示出满意的分离精度。然后使用三名肺栓塞患者的数据对该方法进行测试,定性检查分离图像和放射性核素灌注扫描图像之间的一致性。意义。该方法使用轻量级卷积神经网络进行快速网络训练和推理。它是临床环境中动态心脏相关信号分离的潜在方法。
    Objectives.The cardiac-related component in chest electrical impedance tomography (EIT) measurement is of potential value to pulmonary perfusion monitoring and cardiac function measurement. In a spontaneous breathing case, cardiac-related signals experience serious interference from ventilation-related signals. Traditional cardiac-related signal-separation methods are usually based on certain features of signals. To further improve the separation accuracy, more comprehensive features of the signals should be exploited.Approach.We propose an unsupervised deep-learning method called deep feature-domain matching (DFDM), which exploits the feature-domain similarity of the desired signals and the breath-holding signals. This method is characterized by two sub-steps. In the first step, a novel Siamese network is designed and trained to learn common features of breath-holding signals; in the second step, the Siamese network is used as a feature-matching constraint between the separated signals and the breath-holding signals.Main results.The method is first tested using synthetic data, and the results show satisfactory separation accuracy. The method is then tested using the data of three patients with pulmonary embolism, and the consistency between the separated images and the radionuclide perfusion scanning images is checked qualitatively.Significance.The method uses a lightweight convolutional neural network for fast network training and inference. It is a potential method for dynamic cardiac-related signal separation in clinical settings.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    快速识别植物病害对于有效缓解和控制其对植物的影响至关重要。用于植物病害自动识别,基于深度学习算法的植物叶片图像分类是目前最准确、最流行的方法。现有方法依赖于大量图像标注数据的采集,无法灵活调整识别类别,而我们开发了一种用于自动检测的新图像检索系统,本地化,以及在开放环境中识别单个叶片疾病,即,新增加的疾病类型无需再培训即可识别。在本文中,我们首先优化YOLOv5算法,增强对小物体的识别能力,这有助于更准确地提取叶子对象;其次,将分类识别与度量学习相结合,联合学习对图像进行分类和相似性测量,因此,利用可用图像分类模型的预测能力;最后,构建高效、灵活的图像检索系统,快速确定叶部病害类型。我们在三个公开的叶片疾病数据集上展示了详细的实验结果,并证明了我们系统的有效性。这项工作为促进适用于智能农业和营养诊断等作物研究的植物疾病监测奠定了基础,健康状况监测,还有更多.
    Rapid identification of plant diseases is essential for effective mitigation and control of their influence on plants. For plant disease automatic identification, classification of plant leaf images based on deep learning algorithms is currently the most accurate and popular method. Existing methods rely on the collection of large amounts of image annotation data and cannot flexibly adjust recognition categories, whereas we develop a new image retrieval system for automated detection, localization, and identification of individual leaf disease in an open setting, namely, where newly added disease types can be identified without retraining. In this paper, we first optimize the YOLOv5 algorithm, enhancing recognition ability in small objects, which helps to extract leaf objects more accurately; secondly, integrating classification recognition with metric learning, jointly learning categorizing images and similarity measurements, thus, capitalizing on prediction ability of available image classification models; and finally, constructing an efficient and nimble image retrieval system to quickly determine leaf disease type. We demonstrate detailed experimental results on three publicly available leaf disease datasets and prove the effectiveness of our system. This work lays the groundwork for promoting disease surveillance of plants applicable to intelligent agriculture and to crop research such as nutrition diagnosis, health status surveillance, and more.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    室性早搏(PVC),在普通人群和患者人群中很常见,不规则的心跳表明潜在的心脏病。临床上,从可穿戴设备收集的长期心电图(ECG)是一种非侵入性和廉价的工具,广泛用于医生诊断PVC。然而,对于心脏病学家来说,分析这些长期心电图既费时又费力.因此,本文提出了一种简单但功能强大的从长期心电图中检测PVC的方法。建议的方法利用深度度量学习来提取特征,产品内方差紧凑,产品间差异分离,从心跳。随后,k-最近邻(KNN)分类器根据这些特征计算样本之间的距离以检测PVC。与以前用于检测PVC的系统不同,所提出的过程可以通过有监督的深度度量学习智能自动提取特征,可以避免人工特征工程带来的偏差。作为一套普遍可用的标准测试材料,MIT-BIH(麻省理工学院贝斯以色列医院)心律失常数据库用于评估所提出的方法,实验准确率为99.7%,灵敏度97.45%,和99.87%的特异性。仿真事件表明,使用深度度量学习和KNN进行PVC识别是可靠的。更重要的是,整体方式不依赖于复杂和繁琐的预处理。
    Premature ventricular contractions (PVCs), common in the general and patient population, are irregular heartbeats that indicate potential heart diseases. Clinically, long-term electrocardiograms (ECG) collected from the wearable device is a non-invasive and inexpensive tool widely used to diagnose PVCs by physicians. However, analyzing these long-term ECG is time-consuming and labor-intensive for cardiologists. Therefore, this paper proposed a simplistic but powerful approach to detect PVC from long-term ECG. The suggested method utilized deep metric learning to extract features, with compact intra-product variance and separated inter-product differences, from the heartbeat. Subsequently, the k-nearest neighbors (KNN) classifier calculated the distance between samples based on these features to detect PVC. Unlike previous systems used to detect PVC, the proposed process can intelligently and automatically extract features by supervised deep metric learning, which can avoid the bias caused by manual feature engineering. As a generally available set of standard test material, the MIT-BIH (Massachusetts Institute of Technology-Beth Israel Hospital) Arrhythmia Database is used to evaluate the proposed method, and the experiment takes 99.7% accuracy, 97.45% sensitivity, and 99.87% specificity. The simulation events show that it is reliable to use deep metric learning and KNN for PVC recognition. More importantly, the overall way does not rely on complicated and cumbersome preprocessing.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Sci-hub)

       PDF(Pubmed)

  • 文章类型: Journal Article
    卷积神经网络(CNN)是图像检索领域最主流的解决方案。将深度度量学习引入图像检索领域,重点研究了基于对偶的损失函数的构造。然而,度量学习的大多数基于对的损失函数仅考虑最终图像描述符的公共向量相似性(例如欧几里德距离),而忽略这些描述符的其他分布字符。在这项工作中,我们提出了相对分布熵(RDE)来描述图像描述符的内部分布属性。我们将相对分布熵与欧氏距离相结合,得到相对分布熵加权距离(RDE-distance)。此外,将RDE距离与对比损失和三重态损失融合,建立相对分布熵损失函数。实验结果表明,我们的方法在大多数图像检索基准上都达到了最先进的性能。
    Convolutional neural networks (CNN) is the most mainstream solution in the field of image retrieval. Deep metric learning is introduced into the field of image retrieval, focusing on the construction of pair-based loss function. However, most pair-based loss functions of metric learning merely take common vector similarity (such as Euclidean distance) of the final image descriptors into consideration, while neglecting other distribution characters of these descriptors. In this work, we propose relative distribution entropy (RDE) to describe the internal distribution attributes of image descriptors. We combine relative distribution entropy with the Euclidean distance to obtain the relative distribution entropy weighted distance (RDE-distance). Moreover, the RDE-distance is fused with the contrastive loss and triplet loss to build the relative distributed entropy loss functions. The experimental results demonstrate that our method attains the state-of-the-art performance on most image retrieval benchmarks.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Sci-hub)

       PDF(Pubmed)

  • 文章类型: Journal Article
    为了在样品载玻片查看期间区分模糊的图像,病理学家通常花费大量时间从已确认的类似图像或病例中寻求指导,这是低效的。因此,为了方便病理学家获取与查询图像共享相似内容的图像,已经提出了几种组织病理学图像检索方法。然而,这些方法不能确保合理的相似性度量,其中一些需要大量的注释图像来训练特征提取器来表示图像。受这种情况的驱使,本文提出了第一种基于深度度量学习的组织病理学图像检索方法,并构建了基于混合注意力机制的深度神经网络,在图像类别信息的监督下学习嵌入函数。有了学习的嵌入函数,原始图像被映射到预定义的度量空间,其中来自同一类别的相似图像彼此靠近,,从而可以将度量空间中图像对之间的距离视为图像相似度的合理度量。我们在两个组织病理学图像检索数据集上评估了所提出的方法:我们自己建立的数据集和一个名为KimiaPath24的公共数据集,在该数据集上,所提出的方法在Top-1推荐(Recall@1)中的召回率分别为84.04%和97.89%。此外,进一步的实验证实,所提出的方法可以在训练数据较少的情况下达到与已发布的几种方法相当的性能,这在一定程度上弥补了医学图像注释数据的不足。代码可在https://github.com/easonyang1996/DML_HistoImgRetrieval获得。
    To distinguish ambiguous images during specimen slides viewing, pathologists usually spend lots of time to seek guidance from confirmed similar images or cases, which is inefficient. Therefore, several histopathological image retrieval methods have been proposed for pathologists to easily obtain images sharing similar content with the query images. However, these methods cannot ensure a reasonable similarity metric, and some of them need lots of annotated images to train a feature extractor to represent images. Motivated by this circumstance, we propose the first deep metric learning-based histopathological image retrieval method in this paper and construct a deep neural network based on the mixed attention mechanism to learn an embedding function under the supervision of image category information. With the learned embedding function, original images are mapped into the predefined metric space where similar images from the same category are close to each other, so that the distance between image pairs in the metric space can be regarded as a reasonable metric for image similarity. We evaluate the proposed method on two histopathological image retrieval datasets: our self-established dataset and a public dataset called Kimia Path24, on which the proposed method achieves recall in top-1 recommendation (Recall@1) of 84.04% and 97.89% respectively. Moreover, further experiments confirm that the proposed method can achieve comparable performance to several published methods with less training data, which hedges the shortage of annotated medical image data to some extent. Code is available at https://github.com/easonyang1996/DML_HistoImgRetrieval.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Sci-hub)

公众号