channel attention

频道注意力
  • 文章类型: Journal Article
    医学图像的手动注释对于临床专家来说是耗时的;因此,可靠的自动分割将是处理大型医疗数据集的理想方法。在本文中,我们对妊娠早期超声(US)扫描中两个基本测量的检测和分割感兴趣:Nuchal半透明(NT)和冠部长度(CRL)。形状可能会有很大的变化,胎儿超声扫描中解剖结构的位置或大小。我们提出了一种新方法,即早期妊娠超声CRL和NT分割的密集注意感知网络(DA2Net),依靠强大的注意力机制和密集连接的网络来编码特征大小的变化。我们的结果表明,提出的D2ANet与专家手册注释提供了高像素协议(平均JSC=84.21)。
    Manual annotation of medical images is time consuming for clinical experts; therefore, reliable automatic segmentation would be the ideal way to handle large medical datasets. In this paper, we are interested in detection and segmentation of two fundamental measurements in the first trimester ultrasound (US) scan: Nuchal Translucency (NT) and Crown Rump Length (CRL). There can be a significant variation in the shape, location or size of the anatomical structures in the fetal US scans. We propose a new approach, namely Densely Attentional-Aware Network for First Trimester Ultrasound CRL and NT Segmentation (DA2Net), to encode variation in feature size by relying on the powerful attention mechanism and densely connected networks. Our results show that the proposed D2ANet offers high pixel agreement (mean JSC = 84.21) with expert manual annotations.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    随着自动化恶意软件工具包的发展,网络安全面临不断变化的威胁。尽管基于可视化的恶意软件分析已被证明是一种有效的方法,由于在可视化预处理阶段二进制图像的纹理特征的改变,现有方法难以挑战恶意软件样本,导致性能不佳。此外,为了提高分类精度,现有方法通过设计更深层次的神经网络架构来牺牲预测时间。本文提出了PAFE,一种轻量级和基于可视化的快速恶意软件分类方法。它通过像素填充技术解决了预处理中的纹理特征变化问题,并应用数据增强来克服小样本数据集中的类不平衡的挑战。PAFE结合了多尺度特征融合和信道注意机制,通过模块化设计增强功能表达。广泛的实验结果表明,PAFE在恶意软件变体分类的效率和有效性方面都优于当前最先进的方法,准确率为99.25%,预测时间为10.04ms。
    With the development of automated malware toolkits, cybersecurity faces evolving threats. Although visualization-based malware analysis has proven to be an effective method, existing approaches struggle with challenging malware samples due to alterations in the texture features of binary images during the visualization preprocessing stage, resulting in poor performance. Furthermore, to enhance classification accuracy, existing methods sacrifice prediction time by designing deeper neural network architectures. This paper proposes PAFE, a lightweight and visualization-based rapid malware classification method. It addresses the issue of texture feature variations in preprocessing through pixel-filling techniques and applies data augmentation to overcome the challenges of class imbalance in small sample datasets. PAFE combines multi-scale feature fusion and a channel attention mechanism, enhancing feature expression through modular design. Extensive experimental results demonstrate that PAFE outperforms the current state-of-the-art methods in both efficiency and effectiveness for malware variant classification, achieving an accuracy rate of 99.25 % with a prediction time of 10.04 ms.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    傅里叶重叠显微术(FPM)是一种基于光学原理的显微成像技术。它采用傅立叶光学来分离和组合来自样品的不同光学信息。然而,在成像过程中引入的噪声往往导致重建图像的分辨率较差。本文设计了一种基于残差局部混合网络的方法来提高傅立叶重叠重建图像的质量。通过将通道注意力和空间注意力纳入FPM重建过程,提高了网络重构的效率,减少了重构时间。此外,高斯扩散模型的引入进一步减少了相干伪影,提高了图像重建质量。对比实验结果表明,该网络具有较好的重建质量,在主观观察和客观定量评价方面都优于现有方法。
    Fourier Ptychographic Microscopy (FPM) is a microscopy imaging technique based on optical principles. It employs Fourier optics to separate and combine different optical information from a sample. However, noise introduced during the imaging process often results in poor resolution of the reconstructed image. This article has designed an approach based on a residual local mixture network to improve the quality of Fourier ptychographic reconstruction images. By incorporating channel attention and spatial attention into the FPM reconstruction process, the network enhances the efficiency of the network reconstruction and reduces the reconstruction time. Additionally, the introduction of the Gaussian diffusion model further reduces coherent artifacts and improves image reconstruction quality. Comparative experimental results indicate that this network achieves better reconstruction quality, and outperforming existing methods in both subjective observation and objective quantitative evaluation.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    背景:虚拟现实晕动病(VRMS)是阻碍虚拟现实技术发展的关键问题,而准确检测其发生是解决问题的首要前提。
    目的:在本文中,提出了一种基于多尺度特征相关性的卷积神经网络(CNN)脑电检测模型,用于检测VRMS。
    方法:该模型使用多尺度一维卷积层从多导联EEG数据中提取多尺度时间特征,然后计算提取的多尺度特征在所有引线之间的特征相关性,形成特征相邻矩阵,它将时域特征转换为基于相关性的大脑网络特征,从而加强特征表示。最后,各层的相关特征融合。融合的特征然后被馈送到信道关注模块以过滤信道并使用完全连接的网络对其进行分类。最后,我们招募受试者体验6种不同模式的虚拟过山车场景,并收集任务前后的静息脑电数据对模型进行验证。
    结果:结果表明,精度,该模型对VRMS识别的召回率和F1分数为98.66%,98.65%,98.68%,98.66%,分别。所提出的模型优于当前经典和先进的EEG识别模型。
    结论:表明该模型可用于基于静息状态EEG的VRMS识别。
    BACKGROUND: Virtual reality motion sickness (VRMS) is a key issue hindering the development of virtual reality technology, and accurate detection of its occurrence is the first prerequisite for solving the issue.
    OBJECTIVE: In this paper, a convolutional neural network (CNN) EEG detection model based on multi-scale feature correlation is proposed for detecting VRMS.
    METHODS: The model uses multi-scale 1D convolutional layers to extract multi-scale temporal features from the multi-lead EEG data, and then calculates the feature correlations of the extracted multi-scale features among all the leads to form the feature adjacent matrixes, which converts the time-domain features to correlation-based brain network features, thus strengthen the feature representation. Finally, the correlation features of each layer are fused. The fused features are then fed into the channel attention module to filter the channels and classify them using a fully connected network. Finally, we recruit subjects to experience 6 different modes of virtual roller coaster scenes, and collect resting EEG data before and after the task to verify the model.
    RESULTS: The results show that the accuracy, precision, recall and F1-score of this model for the recognition of VRMS are 98.66 %, 98.65 %, 98.68 %, and 98.66 %, respectively. The proposed model outperforms the current classic and advanced EEG recognition models.
    CONCLUSIONS: It shows that this model can be used for the recognition of VRMS based on the resting state EEG.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    玉米种子是农业生产中必不可少的元素,准确识别其品种和质量对于种植管理至关重要,品种改良,和农产品质量控制。然而,更需要传统的人工分类方法来满足智慧农业的需求。随着深度学习方法在计算机领域的快速发展,我们提出了一个有效的残差网络ERNet来识别高光谱玉米种子。首先,我们使用线性判别分析对高光谱玉米种子图像进行降维处理,以使图像可以平滑地输入到网络中。第二,我们使用有效的残差块从图像中提取细粒度特征。最后,我们使用分类器softmax对高光谱玉米种子图像进行检测和分类。与其他深度学习技术和传统方法相比,ERNet的表现非常出色。准确率为98.36%,该结果为分类研究提供了有价值的参考,包括高光谱玉米种子图片。
    Corn seeds are an essential element in agricultural production, and accurate identification of their varieties and quality is crucial for planting management, variety improvement, and agricultural product quality control. However, more than traditional manual classification methods are needed to meet the needs of intelligent agriculture. With the rapid development of deep learning methods in the computer field, we propose an efficient residual network named ERNet to identify hyperspectral corn seeds. First, we use linear discriminant analysis to perform dimensionality reduction processing on hyperspectral corn seed images so that the images can be smoothly input into the network. Second, we use effective residual blocks to extract fine-grained features from images. Lastly, we detect and categorize the hyperspectral corn seed images using the classifier softmax. ERNet performs exceptionally well compared to other deep learning techniques and conventional methods. With 98.36% accuracy rate, the result is a valuable reference for classification studies, including hyperspectral corn seed pictures.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    最近,无监督算法在图像去雾方面取得了显著的性能。然而,由于数据分布不一致,CycleGAN框架可能会导致生成器学习中的混乱,并且DisentGAN框架对生成的图像缺乏有效的约束,导致图像内容细节丢失和颜色失真。此外,挤压和激励通道注意力仅使用完全连接的层来捕获全局信息,缺乏与当地信息的互动,导致图像去雾的特征权重分配不准确。为了解决上述问题,在本文中,我们提出了一种无监督双向对比重构和自适应细粒度信道注意网络(UBRFC-Net)。具体来说,提出了一种无监督双向对比重建框架(BCRF),旨在建立双向对比重建约束,不仅避免了CycleGAN中的生成器学习混乱,而且增强了对清晰图像的约束能力和无监督去雾网络的重建能力。此外,开发了一种自适应细粒度信道注意力(FCA),以利用相关矩阵来捕获各种粒度的全局和局部信息之间的相关性,促进它们之间的相互作用。实现更有效的特征权重分配。在具有挑战性的基准数据集上的实验结果表明,我们的UBRFC-Net优于最先进的无监督图像去雾方法。本研究成功引入了一种增强的无监督图像去雾方法,解决现有方法的局限性,实现卓越的去雾效果。源代码可在https://github.com/Lose-Code/UBRFC-Net上获得。
    Recently, Unsupervised algorithms has achieved remarkable performance in image dehazing. However, the CycleGAN framework can lead to confusion in generator learning due to inconsistent data distributions, and the DisentGAN framework lacks effective constraints on generated images, resulting in the loss of image content details and color distortion. Moreover, Squeeze and Excitation channel attention employs only fully connected layers to capture global information, lacking interaction with local information, resulting in inaccurate feature weight allocation for image dehazing. To solve the above problems, in this paper, we propose an Unsupervised Bidirectional Contrastive Reconstruction and Adaptive Fine-Grained Channel Attention Networks (UBRFC-Net). Specifically, an Unsupervised Bidirectional Contrastive Reconstruction Framework (BCRF) is proposed, aiming to establish bidirectional contrastive reconstruction constraints, not only to avoid the generator learning confusion in CycleGAN but also to enhance the constraint capability for clear images and the reconstruction ability of the unsupervised dehazing network. Furthermore, an Adaptive Fine-Grained Channel Attention (FCA) is developed to utilize the correlation matrix to capture the correlation between global and local information at various granularities promotes interaction between them, achieving more efficient feature weight assignment. Experimental results on challenging benchmark datasets demonstrate the superiority of our UBRFC-Net over state-of-the-art unsupervised image dehazing methods. This study successfully introduces an enhanced unsupervised image dehazing approach, addressing limitations of existing methods and achieving superior dehazing results. The source code is available at https://github.com/Lose-Code/UBRFC-Net.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    在输入图像的局部窗口内执行卷积操作。因此,卷积神经网络(CNN)擅长获取局部信息。同时,自注意(SA)机制通过计算图像中所有位置的标记之间的相关性来提取特征,这在获取全球信息方面具有优势。因此,这两个模块可以相互补充,提高特征提取能力。一种有效的融合方法是一个值得深入研究的问题。在本文中,我们提出了以U-Net为骨干的CNN和SA并行网络CSAP-UNet。编码器由CNN和Transformer两个并行分支组成,用于从输入图像中提取特征,它考虑了全局依赖关系和本地信息。因为医学图像来自频谱中的某些频带,它们的颜色通道不像自然图像那么均匀。同时,医学分割更加关注图像中的病变区域。注意力融合模块(AFM)将通道注意力和空间注意力串联起来,融合两个分支的输出特征。医学图像分割任务实质上是定位图像中对象的边界。边界增强模块(BEM)设计在所提出的网络的浅层中,以更具体地关注像素级边缘细节。在三个公共数据集上的实验结果验证了CSAP-UNet优于最先进的网络,特别是在ISIC2017数据集上。在Kvasir和CVC-ClinicDB上的跨数据集评估表明,CSAP-UNet具有很强的泛化能力。消融实验也表明了所设计模块的有效性。培训和测试代码可在https://github.com/zhouzhou1201/CSAP-UNet获得。git.
    Convolution operation is performed within a local window of the input image. Therefore, convolutional neural network (CNN) is skilled in obtaining local information. Meanwhile, the self-attention (SA) mechanism extracts features by calculating the correlation between tokens from all positions in the image, which has advantage in obtaining global information. Therefore, the two modules can complement each other to improve feature extraction ability. An effective fusion method is a problem worthy of further study. In this paper, we propose a CNN and SA paralleling network CSAP-UNet with U-Net as backbone. The encoder consists of two parallel branches of CNN and Transformer to extract the feature from the input image, which takes into account both the global dependencies and the local information. Because medical images come from certain frequency bands within the spectrum, their color channels are not as uniform as natural images. Meanwhile, medical segmentation pays more attention to lesion regions in the image. Attention fusion module (AFM) integrates channel attention and spatial attention in series to fuse the output features of the two branches. The medical image segmentation task is essentially to locate the boundary of the object in the image. The boundary enhancement module (BEM) is designed in the shallow layer of the proposed network to focus more specifically on pixel-level edge details. Experimental results on three public datasets validate that CSAP-UNet outperforms state-of-the-art networks, particularly on the ISIC 2017 dataset. The cross-dataset evaluation on Kvasir and CVC-ClinicDB shows that CSAP-UNet has strong generalization ability. Ablation experiments also indicate the effectiveness of the designed modules. The code for training and test is available at https://github.com/zhouzhou1201/CSAP-UNet.git.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    旨在帮助临床医生选择治疗的机器学习方法受到越来越多的关注。因此,这引起了人们对利用脑电图(EEG)数据的癫痫自动检测系统的高度关注。然而,为了使识别模型能够准确地捕获与通道相关的各种特征,频率,和时间信息,有必要具有正确表示的EEG数据。为了应对这一挑战,我们提出了一种基于残差的混合注意力网络(RIHANet)来实现自动癫痫发作检测。最初,通过采用经验模式分解和短时傅里叶变换(EMD-STFT)进行数据处理,提高了EEG的时频表征质量。此外,通过将一种新的基于残差的盗梦空间应用于网络架构,检测模型可以学习局部和全局多尺度时空特征。此外,所设计的混合注意力模型用于从多个角度获取脑电信号之间的关系,包括频道,子空间,和全球。使用四个公共数据集,建议的方法优于最近的学术出版物的结果。
    Increasing attention is being given to machine learning methods designed to aid clinicians in treatment selection. Therefore, this has aroused a heightened focus on the auto-detect system of epilepsy utilizing electroencephalogram(EEG) data. However, in order for the recognition model to accurately capture a wide range of features related to channel, frequency, and temporal information, it is necessary to have EEG data that is correctly represented. To tackle this challenge, we propose a Residual-based Inception with Hybrid-Attention Network(RIHANet) to achieve automatic seizure detection. Initially, by employing Empirical Mode Decomposition and Short-time Fourier Transform(EMD-STFT) for data processing, it can improve the quality of time-frequency representation of EEG. Additionally, by applying a novel Residual-based Inception to the network architecture, the detection model can learn local and global multiscale spatial-temporal features. Furthermore, the Hybrid Attention model designed is used to obtain relationships between EEG signals from multiple perspectives, including channels, sub-spaces, and global. Using four public datasets, the suggested approach outperforms the results in the most recent scholarly publications.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    目的:小肠溃疡,发病率很高,包括克罗恩病(CD),肠结核(ITB),原发性小肠淋巴瘤(PSIL),隐源性多灶性溃疡性狭窄性肠炎(CMUSE),和非特异性溃疡(NSU)。然而,肠镜检查易误诊溃疡形态。
    方法:在本研究中,DRCA-DenseNet169,它基于DenseNet169,具有残余扩张块和通道注意块,建议识别CD,ITB,PSIL,CMUSE,和NSU智能。此外,一个包含动态权重的新型损失函数旨在提高有限样本不平衡数据集的精度。使用10883张肠镜检查图像评估DRCA-Densenet169,包括5375张溃疡图像和5508张正常图像,从上海长海医院获得。
    结果:DRCA-Densenet169的总体准确度为85.27%±0.32%,加权精度为83.99%±2.47%,加权召回率为84.36%±0.88%,加权F1评分为84.07%±2.14%。
    结论:结果表明,在获得即时和初步诊断时,DRCA-Densenet169在识别不同类型的溃疡方面具有很高的识别准确性和很强的鲁棒性。
    Objective. Ulceration of the small intestine, which has a high incidence, includes Crohn\'s disease (CD), intestinal tuberculosis (ITB), primary small intestinal lymphoma (PSIL), cryptogenic multifocal ulcerous stenosing enteritis (CMUSE), and non-specific ulcer (NSU). However, the ulceration morphology can easily be misdiagnosed through enteroscopy.Approach. In this study, DRCA-DenseNet169, which is based on DenseNet169, with residual dilated blocks and a channel attention block, is proposed to identify CD, ITB, PSIL, CMUSE, and NSU intelligently. In addition, a novel loss function that incorporates dynamic weights is designed to enhance the precision of imbalanced datasets with limited samples. DRCA-Densenet169 was evaluated using 10883 enteroscopy images, including 5375 ulcer images and 5508 normal images, which were obtained from the Shanghai Changhai Hospital.Main results. DRCA-Densenet169 achieved an overall accuracy of 85.27% ± 0.32%, a weighted-precision of 83.99% ± 2.47%, a weighted-recall of 84.36% ± 0.88% and a weighted-F1-score of 84.07% ± 2.14%.Significance. The results demonstrate that DRCA-Densenet169 has high recognition accuracy and strong robustness in identifying different types of ulcers when obtaining immediate and preliminary diagnoses.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    高血压视网膜病变(HR)是由高血压引发的微血管视网膜变化引起的,这是全球可预防失明最常见的主要原因。因此,有必要开发一种使用视网膜图像进行HR检测和评估的自动化系统。我们旨在提出一种自动方法来识别和分类不同程度的HR严重程度。一种称为空间卷积模块(SCM)的新网络结合了跨通道和空间信息,和卷积运算提取有用的特征。本模型使用可公开访问的数据集ODIR进行评估,INSPIREVR,和VICAVR。我们应用增强来人为地增加1200张眼底图像的数据集。正常的不同HR严重程度,温和,中度,严重,与现有模型相比,恶性和恶性最终被归类为减少的时间,因为在提出的模型中,卷积层在输入眼底图像上只运行一次,这导致加速并减少检测血管结构异常的处理时间。根据调查结果,改进的SVM在血管分类中具有最高的检测和分类准确率,准确率为98.99%,在160.4s内完成任务。即,比5倍分类准确率高0.27,改进的KNN分类器实现了98.72%的准确率。当计算效率优先时,所提出的模型快速识别不同HR严重程度的能力是显著的。
    Hypertensive retinopathy (HR) results from the microvascular retinal changes triggered by hypertension, which is the most common leading cause of preventable blindness worldwide. Therefore, it is necessary to develop an automated system for HR detection and evaluation using retinal images. We aimed to propose an automated approach to identify and categorize the various degrees of HR severity. A new network called the spatial convolution module (SCM) combines cross-channel and spatial information, and the convolution operations extract helpful features. The present model is evaluated using publicly accessible datasets ODIR, INSPIREVR, and VICAVR. We applied the augmentation to artificially increase the dataset of 1200 fundus images. The different HR severity levels of normal, mild, moderate, severe, and malignant are finally classified with the reduced time when compared to the existing models because in the proposed model, convolutional layers run only once on the input fundus images, which leads to a speedup and reduces the processing time in detecting the abnormalities in the vascular structure. According to the findings, the improved SVM had the highest detection and classification accuracy rate in the vessel classification with an accuracy of 98.99% and completed the task in 160.4 s. The ten-fold classification achieved the highest accuracy of 98.99%, i.e., 0.27 higher than the five-fold classification accuracy and the improved KNN classifier achieved an accuracy of 98.72%. When computation efficiency is a priority, the proposed model\'s ability to quickly recognize different HR severity levels is significant.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

公众号