Multi-head attention

多头注意力
  • 文章类型: Journal Article
    Objective.深度学习显着增强了稀疏视图计算机断层扫描重建的性能。然而,这些方法对使用高质量配对数据集的监督训练的依赖性,以及在各种身体获取条件下进行再培训的必要性,限制它们在新的成像环境和设置中的通用性。方法。为了克服这些限制,我们提出了一种基于深度图像先验框架的无监督方法。我们的方法通过结合多级线性扩散噪声,超越了传统的单噪声级输入,显着降低过度拟合的风险。此外,我们将非局部自相似性作为深度隐式先验嵌入到自我注意力网络结构中,提高模型识别和利用整个图像重复模式的能力。此外,利用成像物理学,在图像域和投影数据空间之间进行梯度反向传播以优化网络权重。主要结果。模拟和临床病例的评估证明了我们的方法在各种投影视图中的有效零射适应性,突出其鲁棒性和灵活性。此外,我们的方法有效地消除了噪声和条纹伪影,同时显着恢复复杂的图像细节。意义。我们的方法旨在克服当前基于监督深度学习的稀疏视图CT重建的局限性,提供改进的泛化性和适应性,而不需要大量的成对训练数据。
    Objective.Deep learning has markedly enhanced the performance of sparse-view computed tomography reconstruction. However, the dependence of these methods on supervised training using high-quality paired datasets, and the necessity for retraining under varied physical acquisition conditions, constrain their generalizability across new imaging contexts and settings.Approach.To overcome these limitations, we propose an unsupervised approach grounded in the deep image prior framework. Our approach advances beyond the conventional single noise level input by incorporating multi-level linear diffusion noise, significantly mitigating the risk of overfitting. Furthermore, we embed non-local self-similarity as a deep implicit prior within a self-attention network structure, improving the model\'s capability to identify and utilize repetitive patterns throughout the image. Additionally, leveraging imaging physics, gradient backpropagation is performed between the image domain and projection data space to optimize network weights.Main Results.Evaluations with both simulated and clinical cases demonstrate our method\'s effective zero-shot adaptability across various projection views, highlighting its robustness and flexibility. Additionally, our approach effectively eliminates noise and streak artifacts while significantly restoring intricate image details.Significance. Our method aims to overcome the limitations in current supervised deep learning-based sparse-view CT reconstruction, offering improved generalizability and adaptability without the need for extensive paired training data.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    准确预测锂离子电池的健康状态(SOH)是估计其剩余寿命的基础。各种参数,如电压,电流,和温度显著影响电池的SOH。然而,现有的数据驱动方法需要来自目标领域的大量数据进行训练,这阻碍了在初始阶段对锂离子电池健康状况的评估。为了应对这些挑战,本文介绍了多头注意时间卷积网络(MHAT-TCN),将多头注意力学习与随机块退出技术结合起来。此外,它采用灰色关联分析(GRA)来选择与电池容量高度相关的健康指标(HIs),从而提高模型训练的准确性。采用留空交叉验证(LOOCV),MHAT-TCN网络使用来自相同模型的电池的数据进行预训练,以便于在整个操作期间对目标电池进行综合预测。结果表明,在HIs上训练的MHAT-TCN网络优于其他模型,在整个运营期间实现精确预测。
    Accurately predicting the state of health (SOH) of lithium-ion batteries is fundamental in estimating their remaining lifespan. Various parameters such as voltage, current, and temperature significantly influence the battery\'s SOH. However, existing data-driven methods necessitate substantial data from the target domain for training, which hampers the assessment of lithium-ion battery health at the initial stage. To address these challenges, this paper introduces the multi-head attention-time convolution network (MHAT-TCN), amalgamating multi-head attention learning with random block dropout techniques. Additionally, it employs grey relational analysis (GRA) to select health indicators (HIs) highly correlated with battery capacity, thereby enhancing the accuracy of the model training. Employing leave-one-out crossvalidation (LOOCV), the MHAT-TCN network is pre-trained using data from batteries of the same model to facilitate comprehensive prediction of the target battery throughout its operational period. Results demonstrate that the MHAT-TCN network trained on HIs outperforms other models, enabling precise predictions across the entire operational period.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    黑色素瘤是一种皮肤癌,具有重大的健康风险,需要早期发现才能有效治疗。本研究提出了一种新颖的方法,该方法将基于变压器的模型与手工制作的纹理特征和灰狼优化集成在一起,旨在提高黑色素瘤分类的效率。预处理涉及标准化图像尺寸和通过中值滤波技术提高图像质量。纹理特征,包括GLCM和LBP,被提取以捕获指示黑素瘤的空间模式。GWO算法用于选择最有区别的特征。然后采用基于变压器的解码器进行分类,利用注意力机制来捕获上下文依赖关系。在HAM10000数据集和ISIC2019数据集上的实验验证显示了所提出方法的有效性。基于变压器的模型,结合手工制作的纹理特征,并在灰狼优化的指导下,取得了突出的成绩。结果表明,该方法在黑色素瘤检测任务中表现良好,在HAM10000数据集上实现99.54%和99.11%的准确性和F1得分,准确率为99.47%,ISIC2019数据集上的F1评分为99.25%。•我们使用LBP和GLCM的概念从皮肤病变图像中提取特征。•灰狼优化(GWO)算法用于特征选择。•基于变形金刚的解码器用于黑素瘤分类。
    Melanoma is a type of skin cancer that poses significant health risks and requires early detection for effective treatment. This study proposing a novel approach that integrates a transformer-based model with hand-crafted texture features and Gray Wolf Optimization, aiming to enhance efficiency of melanoma classification. Preprocessing involves standardizing image dimensions and enhancing image quality through median filtering techniques. Texture features, including GLCM and LBP, are extracted to capture spatial patterns indicative of melanoma. The GWO algorithm is applied to select the most discriminative features. A transformer-based decoder is then employed for classification, leveraging attention mechanisms to capture contextual dependencies. The experimental validation on the HAM10000 dataset and ISIC2019 dataset showcases the effectiveness of the proposed methodology. The transformer-based model, integrated with hand-crafted texture features and guided by Gray Wolf Optimization, achieves outstanding results. The results showed that the proposed method performed well in melanoma detection tasks, achieving an accuracy and F1-score of 99.54% and 99.11% on the HAM10000 dataset, and an accuracy of 99.47%, and F1-score of 99.25% on the ISIC2019 dataset. • We use the concepts of LBP and GLCM to extract features from the skin lesion images. • The Gray Wolf Optimization (GWO) algorithm is employed for feature selection. • A decoder based on Transformers is utilized for melanoma classification.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    背景:自闭症谱系障碍(ASD)是一种神经发育障碍,在患者中表现出异质性特征,包括发育进展的变异性和受性别和年龄影响的独特神经解剖学特征。基于功能连接(FC)图的深度学习模型的最新进展产生了有希望的结果,但是他们专注于普遍的全球激活模式,未能捕获专门的区域特征并准确评估疾病适应症。
    方法:为了克服这些限制,我们提出了一种新颖的深度学习方法,该方法对具有多头注意力的FC进行建模,它可以同时对与ASD相关的复杂和可变的大脑连接模式进行建模,有效提取大脑连接的异常模式。所提出的方法不仅识别特定区域的相关性,而且强调特定区域的连接,从不同的角度来看,瞬态时间点。提取的FC被转换为图形,将加权标签分配给边缘以反映相关程度,然后使用能够处理边缘标签的图神经网络进行处理。
    结果:关于自闭症脑成像数据交换(ABIDE)I和II数据集的实验,其中包括一个异质的队列,表现出优于最先进的方法,提高精度高达3.7%。在FC分析中加入多头注意力显着改善了典型大脑与受ASD影响的大脑之间的区别。此外,消融研究验证了不同年龄和性别的ASD患者的不同大脑特征,提供有见地的解释。
    结论:这些结果强调了该方法在提高诊断准确性方面的有效性及其在推进ASD诊断的神经学研究方面的潜力。
    BACKGROUND: Autism spectrum disorder (ASD) is a neurodevelopmental disorder exhibiting heterogeneous characteristics in patients, including variability in developmental progression and distinct neuroanatomical features influenced by sex and age. Recent advances in deep learning models based on functional connectivity (FC) graphs have produced promising results, but they have focused on generalized global activation patterns and failed to capture specialized regional characteristics and accurately assess disease indications.
    METHODS: To overcome these limitations, we propose a novel deep learning method that models FC with multi-head attention, which enables simultaneous modeling of the intricate and variable patterns of brain connectivity associated with ASD, effectively extracting abnormal patterns of brain connectivity. The proposed method not only identifies region-specific correlations but also emphasizes connections at specific, transient time points from diverse perspectives. The extracted FC is transformed into a graph, assigning weighted labels to the edges to reflect the degree of correlation, which is then processed using a graph neural network capable of handling edge labels.
    RESULTS: Experiments on the autism brain imaging data exchange (ABIDE) I and II datasets, which include a heterogeneous cohort, showed superior performance over the state-of-the-art methods, improving accuracy by up to 3.7%p. The incorporation of multi-head attention in FC analysis markedly improved the distinction between typical brains and those affected by ASD. Additionally, the ablation study validated diverse brain characteristics in ASD patients across different ages and sexes, offering insightful interpretations.
    CONCLUSIONS: These results emphasize the effectiveness of the method in enhancing diagnostic accuracy and its potential in advancing neurological research for ASD diagnosis.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    通过在许多细胞过程中参与高达40%的蛋白质-蛋白质相互作用,肽在许多生物活性中是关键的。由于其特殊的特异性和有效性,肽已成为有希望的候选药物设计。然而,准确预测蛋白质-肽结合亲和力仍然是一个挑战。针对这个问题,我们开发了一种基于卷积神经网络和多头注意力的预测模型PepPAP,这完全依赖于序列特征。这些特征包括物理化学性质,内在紊乱,序列编码,尤其是从16,689种非冗余蛋白质-肽复合物中提取的界面倾向。值得注意的是,我们以前的工作中提出的采用的回归分层交叉验证方案有利于改善具有极端结合亲和力值的情况下的预测。在三个基准测试数据集上:T100,一系列靶向PDZ结构域和CXCR4的肽,PepPAP显示出优异的性能,优于现有方法,并展示了其良好的泛化能力。此外,PepPAP在二元相互作用预测中具有良好的效果,对特征空间分布可视化的分析突出了PepPAP的有效性。据我们所知,PepPAP是第一个基于序列的深度关注模型,用于广泛基因组蛋白质-肽结合亲和力预测,并有可能为基于肽的药物设计提供有价值的见解。
    Peptides are pivotal in numerous biological activities by engaging in up to 40 % of protein-protein interactions in many cellular processes. Due to their exceptional specificity and effectiveness, peptides have emerged as promising candidates for drug design. However, accurately predicting protein-peptide binding affinity remains a challenging. Aiming at the problem, we develop a prediction model PepPAP based on convolutional neural network and multi-head attention, which relies solely on sequence features. These features include physicochemical properties, intrinsic disorder, sequence encoding, and especially interface propensity which is extracted from 16,689 non-redundant protein-peptide complexes. Notably, the adopted regression stratification cross-validation scheme proposed in our previous work is beneficial to improve the prediction for the cases with extreme binding affinity values. On three benchmark test datasets: T100, a series of peptides targeting to PDZ domain and CXCR4, PepPAP shows excellent performance, outperforming the existing methods and demonstrating its good generalization ability. Furthermore, PepPAP has good results in binary interaction prediction, and the analysis of the feature space distribution visualization highlights PepPAP\'s effectiveness. To the best of our knowledge, PepPAP is the first sequence-based deep attention model for wide-genome protein-peptide binding affinity prediction, and holds the potential to offer valuable insights for the peptide-based drug design.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    人体液中的分泌蛋白具有作为疾病生物标志物的潜力。这些生物标志物可用于疾病的早期诊断和风险预测。因此对人体体液分泌蛋白的研究具有很大的应用价值。近年来,基于深度学习的转换语言模型已经从自然语言处理(NLP)领域转移到蛋白质组学领域,导致用于蛋白质序列表示的蛋白质语言模型(PLMs)的发展。这里,我们提出了一个名为ESM预测分泌蛋白质(ESMSec)的深度学习框架,以预测人体体液中分泌的三种蛋白质。ESMSec基于ESM2模型和注意力架构。具体来说,首先将蛋白质序列数据放入ESM2模型,从最后一个隐藏层提取特征信息,所有的输入蛋白质都被编码成一个固定的1000×480矩阵。其次,采用具有完全连接的神经网络的多头注意力作为分类器,根据它们是否被分泌到每个体液中进行二元分类。我们的实验利用了三种人体体液,它们是重要且普遍存在的标志物。实验结果表明,EMSec在等离子体测试数据集上达到了0.8486、0.8358和0.8325的平均精度,脑脊液(CSF),和精液,平均而言,它的表现优于最先进的(SOTA)方法。ESMSec的出色性能结果表明,ESM可以提高模型的预测性能,并且在筛选人体体液蛋白质的分泌信息方面具有很大的潜力。
    The secreted proteins of human body fluid have the potential to be used as biomarkers for diseases. These biomarkers can be used for early diagnosis and risk prediction of diseases, so the study of secreted proteins of human body fluid has great application value. In recent years, the deep-learning-based transformer language model has transferred from the field of natural language processing (NLP) to the field of proteomics, leading to the development of protein language models (PLMs) for protein sequence representation. Here, we propose a deep learning framework called ESM Predict Secreted Proteins (ESMSec) to predict three types of proteins secreted in human body fluid. The ESMSec is based on the ESM2 model and attention architecture. Specifically, the protein sequence data are firstly put into the ESM2 model to extract the feature information from the last hidden layer, and all the input proteins are encoded into a fixed 1000 × 480 matrix. Secondly, multi-head attention with a fully connected neural network is employed as the classifier to perform binary classification according to whether they are secreted into each body fluid. Our experiment utilized three human body fluids that are important and ubiquitous markers. Experimental results show that ESMSec achieved average accuracy of 0.8486, 0.8358, and 0.8325 on the testing datasets for plasma, cerebrospinal fluid (CSF), and seminal fluid, which on average outperform the state-of-the-art (SOTA) methods. The outstanding performance results of ESMSec demonstrate that the ESM can improve the prediction performance of the model and has great potential to screen the secretion information of human body fluid proteins.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    背景:准确识别联合用药的风险水平对于研究联合用药的机制和不良反应具有重要意义。大多数现有方法只能预测两种药物之间是否存在相互作用,但不能直接确定其准确的风险水平。
    方法:在本研究中,我们提出了一种名为AERGCN-DDI的多类药物组合风险预测模型,利用具有多头注意机制的关系图卷积网络。具有不同风险水平的药物-药物相互作用事件被建模为异构信息图。基于化合物化学结构信息学习药物节点和链接的属性特征。最后,提出了基于异构图神经网络和多头注意模块的AERGCN-DDI模型预测药物组合风险水平。
    结果:为了评估所提出方法的有效性,我们进行了5倍交叉验证和消融研究.此外,我们将其预测性能与基线模型和其他最先进的方法在两个基准数据集上进行了比较.实证研究证明了AERGCN-DDI的优异性能。
    结论:AERGCN-DDI成为预测药物组合风险水平的有价值的工具,从而帮助临床用药决策,减轻严重的药物副作用,提高患者的临床预后。
    BACKGROUND: Accurately identifying the risk level of drug combinations is of great significance in investigating the mechanisms of combination medication and adverse reactions. Most existing methods can only predict whether there is an interaction between two drugs, but cannot directly determine their accurate risk level.
    METHODS: In this study, we propose a multi-class drug combination risk prediction model named AERGCN-DDI, utilizing a relational graph convolutional network with a multi-head attention mechanism. Drug-drug interaction events with varying risk levels are modeled as a heterogeneous information graph. Attribute features of drug nodes and links are learned based on compound chemical structure information. Finally, the AERGCN-DDI model is proposed to predict drug combination risk level based on heterogenous graph neural network and multi-head attention modules.
    RESULTS: To evaluate the effectiveness of the proposed method, five-fold cross-validation and ablation study were conducted. Furthermore, we compared its predictive performance with baseline models and other state-of-the-art methods on two benchmark datasets. Empirical studies demonstrated the superior performances of AERGCN-DDI.
    CONCLUSIONS: AERGCN-DDI emerges as a valuable tool for predicting the risk levels of drug combinations, thereby aiding in clinical medication decision-making, mitigating severe drug side effects, and enhancing patient clinical prognosis.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    背景:药物-药物相互作用事件影响药物组合的有效性,并可能导致意想不到的副作用或加剧潜在疾病,危及患者预后。大多数现有方法仅限于预测两种药物是否相互作用或药物相互作用的类型。虽然很少有研究试图预测药物组合副作用的具体风险水平。
    方法:在本研究中,我们提议MathEagle,一种基于多头注意力和异质属性图学习的药物组合风险预测方法。最初,我们将药物和药物之间的三种不同风险水平建模为异质信息图.随后,通过消息传递神经网络和图嵌入算法利用药物的行为和化学结构特征,分别。最终,MathEagle采用异构图卷积和多头注意力机制来学习药物节点的有效潜在表示,并以端到端方式估计成对药物的风险水平。
    结果:为了评估模型的有效性和鲁棒性,五折交叉验证,消融实验,并进行了案例研究。MathEagle在药物风险水平预测任务上实现了85.85%的准确度和0.9701的AUC,并且优于所有比较模型。MathEagle预测器可在http://120.77.11.78/MathEagle/免费访问。
    结论:实验结果表明,MathEagle可以作为预测药物组合的准确风险的有效工具,协助指导临床用药,并提高患者的治疗效果。
    BACKGROUND: Drug-drug interaction events influence the effectiveness of drug combinations and can lead to unexpected side effects or exacerbate underlying diseases, jeopardizing patient prognosis. Most existing methods are restricted to predicting whether two drugs interact or the type of drug-drug interactions, while very few studies endeavor to predict the specific risk levels of side effects of drug combinations.
    METHODS: In this study, we propose MathEagle, a novel approach to predict accurate risk levels of drug combinations based on multi-head attention and heterogeneous attribute graph learning. Initially, we model drugs and three distinct risk levels between drugs as a heterogeneous information graph. Subsequently, behavioral and chemical structure features of drugs are utilized by message passing neural networks and graph embedding algorithms, respectively. Ultimately, MathEagle employs heterogeneous graph convolution and multi-head attention mechanisms to learn efficient latent representations of drug nodes and estimates the risk levels of pairwise drugs in an end-to-end manner.
    RESULTS: To assess the effectiveness and robustness of the model, five-fold cross-validation, ablation experiments, and case studies were conducted. MathEagle achieved an accuracy of 85.85 % and an AUC of 0.9701 on the drug risk level prediction task and is superior to all comparative models. The MathEagle predictor is freely accessible at http://120.77.11.78/MathEagle/.
    CONCLUSIONS: The experimental results indicate that MathEagle can function as an effective tool for predicting accurate risk of drug combinations, aiding in guiding clinical medication, and enhancing patient outcomes.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    检测路面裂缝是确保道路安全的重要组成部分。由于手动识别这些裂缝可能很耗时,需要一种自动化的方法来加速这个过程。然而,由于包括裂纹可变性在内的因素,创建这样的系统是具有挑战性的,路面材料的变化,以及路面上杂物和异常现象的发生。受深度学习应用于计算机视觉的最新进展的推动,我们提出了一个有效的U形网模型,称为DepthCrackNet。我们的模型采用了双卷积编码器(DCE),由一系列卷积层组成,用于鲁棒特征提取,同时保持参数最优有效。我们已经将TriInput多头空间注意力(TMSA)模块纳入我们的模型;在这个模块中,每个头独立运行,捕获各种空间关系,促进丰富的上下文信息的提取。此外,DepthCrackNet采用空间深度增强器(SDE)模块,专门设计用于增强我们的分割模型的特征提取能力。在两个公共裂纹数据集上评估了DepthCrackNet的性能:Crack500和DeepCrack。在我们的实验研究中,在Crack500和DeepCrack数据集的情况下,网络的mIoU得分分别为77.0%和83.9%,分别。
    Detecting cracks in the pavement is a vital component of ensuring road safety. Since manual identification of these cracks can be time-consuming, an automated method is needed to speed up this process. However, creating such a system is challenging due to factors including crack variability, variations in pavement materials, and the occurrence of miscellaneous objects and anomalies on the pavement. Motivated by the latest progress in deep learning applied to computer vision, we propose an effective U-Net-shaped model named DepthCrackNet. Our model employs the Double Convolution Encoder (DCE), composed of a sequence of convolution layers, for robust feature extraction while keeping parameters optimally efficient. We have incorporated the TriInput Multi-Head Spatial Attention (TMSA) module into our model; in this module, each head operates independently, capturing various spatial relationships and boosting the extraction of rich contextual information. Furthermore, DepthCrackNet employs the Spatial Depth Enhancer (SDE) module, specifically designed to augment the feature extraction capabilities of our segmentation model. The performance of the DepthCrackNet was evaluated on two public crack datasets: Crack500 and DeepCrack. In our experimental studies, the network achieved mIoU scores of 77.0% and 83.9% with the Crack500 and DeepCrack datasets, respectively.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    从自然场景图像中准确,快速地识别车牌是一项至关重要且具有挑战性的任务。现有的方法可以在简单的场景中识别车牌,但是在复杂的环境中,它们的性能会显著下降。提出了一种新的车牌检测与识别模型YOLOv5-PDLPR,在车牌检测部分采用YOLOv5目标检测算法,在车牌识别部分采用本文提出的PDLPR算法。PDLPR算法主要设计如下:(1)采用多注意机制对单个字符进行准确识别。(2)设计了一种全局特征提取网络,以提高特征提取网络的完备性。(3)采用最新的并行译码器结构,提高推理效率。实验结果表明,该算法比对比算法具有更好的准确率和速度,可以实现实时识别,在复杂场景中具有较高的效率和鲁棒性。
    Accurate and fast recognition of vehicle license plates from natural scene images is a crucial and challenging task. Existing methods can recognize license plates in simple scenarios, but their performance degrades significantly in complex environments. A novel license plate detection and recognition model YOLOv5-PDLPR is proposed, which employs YOLOv5 target detection algorithm in the license plate detection part and uses the PDLPR algorithm proposed in this paper in the license plate recognition part. The PDLPR algorithm is mainly designed as follows: (1) A Multi-Head Attention mechanism is used to accurately recognize individual characters. (2) A global feature extractor network is designed to improve the completeness of the network for feature extraction. (3) The latest parallel decoder architecture is adopted to improve the inference efficiency. The experimental results show that the proposed algorithm has better accuracy and speed than the comparison algorithms, can achieve real-time recognition, and has high efficiency and robustness in complex scenes.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

公众号