transfer learning

迁移学习
  • 文章类型: Journal Article
    确定左心室射血分数(EF)的患者,要么降低[EF<40%(rEF)],中档[EF40-50%(mEF)],或保留[EF>50%(pEF)],被认为是首要的临床重要性。使用GoogleVertexAI中的AutoML的端到端视频分类应用于超声心动图记录。通过多数欠采样平衡的数据集,每个对应于三个可能分类中的一个,是从StandfordEchoNet-Dynamic存储库中获得的。应用75/25的列车测试分裂。rEF与rEF的二元视频分类非rEF表现良好(测试数据集:ROCAUC评分0.939,准确性0.863,敏感性0.894,特异性0.831,阳性预测值0.842)。非pEF与pEF的第二个二元分类pEF表现稍差(测试数据集:ROCAUC评分0.917,准确性0.829,敏感性0.761,特异性0.891,阳性预测值0.888)。还探索了三元分类,并且观察到较低的性能,主要是mEF类。开放访问中的非AutoMLPyTorch实现证实了我们方法的可行性。有了这个概念证明,基于迁移学习对EF进行分类的端到端视频分类,以便在前瞻性临床研究中进一步评估。
    Identifying patients with left ventricular ejection fraction (EF), either reduced [EF < 40% (rEF)], mid-range [EF 40-50% (mEF)], or preserved [EF > 50% (pEF)], is considered of primary clinical importance. An end-to-end video classification using AutoML in Google Vertex AI was applied to echocardiographic recordings. Datasets balanced by majority undersampling, each corresponding to one out of three possible classifications, were obtained from the Standford EchoNet-Dynamic repository. A train-test split of 75/25 was applied. A binary video classification of rEF vs. not rEF demonstrated good performance (test dataset: ROC AUC score 0.939, accuracy 0.863, sensitivity 0.894, specificity 0.831, positive predicting value 0.842). A second binary classification of not pEF vs. pEF was slightly less performing (test dataset: ROC AUC score 0.917, accuracy 0.829, sensitivity 0.761, specificity 0.891, positive predicting value 0.888). A ternary classification was also explored, and lower performance was observed, mainly for the mEF class. A non-AutoML PyTorch implementation in open access confirmed the feasibility of our approach. With this proof of concept, end-to-end video classification based on transfer learning to categorize EF merits consideration for further evaluation in prospective clinical studies.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    从组织病理学图像诊断乳腺癌通常是耗时的,并且容易出现人为错误。影响治疗和预后。深度学习诊断方法为提高乳腺癌检测和分类的准确性和效率提供了潜力。然而,他们在有限的数据和癌症类型之间的微妙变化中挣扎。注意机制提供了在克服此类挑战方面显示出希望的特征细化能力。为此,本文提出了高效信道空间注意力网络(ECSAnet),基于EfficientNetV2构建的架构,并使用卷积块注意力模块(CBAM)和其他完全连接的层进行增强。ECSAnet使用BreakHis数据集进行了微调,采用Reinhardstain归一化和图像增强技术,以最大程度地减少过拟合并增强泛化性。在测试中,ECSAnet的表现优于AlexNet,DenseNet121,EfficientNetV2-S,在大多数设置中,InceptionNetV3、ResNet50和VGG16,在40×时达到94.2%的精度,100×时92.96%,200×88.41%,400倍放大倍数为89.42%。结果强调了CBAM在提高分类准确性方面的有效性以及染色标准化对可泛化性的重要性。
    Breast cancer diagnosis from histopathology images is often time consuming and prone to human error, impacting treatment and prognosis. Deep learning diagnostic methods offer the potential for improved accuracy and efficiency in breast cancer detection and classification. However, they struggle with limited data and subtle variations within and between cancer types. Attention mechanisms provide feature refinement capabilities that have shown promise in overcoming such challenges. To this end, this paper proposes the Efficient Channel Spatial Attention Network (ECSAnet), an architecture built on EfficientNetV2 and augmented with a convolutional block attention module (CBAM) and additional fully connected layers. ECSAnet was fine-tuned using the BreakHis dataset, employing Reinhard stain normalization and image augmentation techniques to minimize overfitting and enhance generalizability. In testing, ECSAnet outperformed AlexNet, DenseNet121, EfficientNetV2-S, InceptionNetV3, ResNet50, and VGG16 in most settings, achieving accuracies of 94.2% at 40×, 92.96% at 100×, 88.41% at 200×, and 89.42% at 400× magnifications. The results highlight the effectiveness of CBAM in improving classification accuracy and the importance of stain normalization for generalizability.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    急性淋巴细胞白血病,通常被称为所有,是一种可以影响血液和骨髓的癌症。诊断过程是一个困难的过程,因为它经常需要专家测试,比如验血,骨髓穿刺,还有活检,所有这些都非常耗时和昂贵。必须获得ALL的早期诊断,以便及时和适当地开始治疗。在最近的医学诊断中,人工智能(AI)和物联网(IoT)设备的集成取得了实质性进展。我们的提案引入了一种新的基于AI的医疗物联网(IoMT)框架,旨在从外周血涂片(PBS)图像中自动识别白血病。在这项研究中,我们提出了一种新的基于深度学习的融合模型来检测所有类型的白血病。系统将诊断报告无缝地提供给集中式数据库,包括患者特定的设备。从医院采集血样后,PBS图像通过支持WiFi的微观设备传输到云服务器。在云服务器中,配置了能够对PBS图像中的ALL进行分类的新融合模型。使用包括来自89个个体的6512个原始和分割图像的数据集来训练融合模型。在融合模型中,两个输入通道用于特征提取。这些通道包括原始图像和分割图像。VGG16负责从原始图像中提取特征,而DenseNet-121负责从分割图像中提取特征。两个输出特征合并在一起,和致密层用于白血病的分类。已经提出的融合模型获得了99.89%的准确率,精度为99.80%,召回率达到99.72%,这使它在白血病分类中处于很好的位置。所提出的模型在性能方面优于几种最先进的卷积神经网络(CNN)模型。因此,这个提出的模型有可能挽救生命和努力。为了更全面地模拟整个方法,本研究开发了一个网络应用程序(测试版)。本申请旨在确定个体中是否存在白血病。这项研究的结果具有在生物医学研究中应用的巨大潜力,特别是提高计算机辅助白血病检测的准确性。
    Acute lymphoblastic leukemia, commonly referred to as ALL, is a type of cancer that can affect both the blood and the bone marrow. The process of diagnosis is a difficult one since it often calls for specialist testing, such as blood tests, bone marrow aspiration, and biopsy, all of which are highly time-consuming and expensive. It is essential to obtain an early diagnosis of ALL in order to start therapy in a timely and suitable manner. In recent medical diagnostics, substantial progress has been achieved through the integration of artificial intelligence (AI) and Internet of Things (IoT) devices. Our proposal introduces a new AI-based Internet of Medical Things (IoMT) framework designed to automatically identify leukemia from peripheral blood smear (PBS) images. In this study, we present a novel deep learning-based fusion model to detect ALL types of leukemia. The system seamlessly delivers the diagnostic reports to the centralized database, inclusive of patient-specific devices. After collecting blood samples from the hospital, the PBS images are transmitted to the cloud server through a WiFi-enabled microscopic device. In the cloud server, a new fusion model that is capable of classifying ALL from PBS images is configured. The fusion model is trained using a dataset including 6512 original and segmented images from 89 individuals. Two input channels are used for the purpose of feature extraction in the fusion model. These channels include both the original and the segmented images. VGG16 is responsible for extracting features from the original images, whereas DenseNet-121 is responsible for extracting features from the segmented images. The two output features are merged together, and dense layers are used for the categorization of leukemia. The fusion model that has been suggested obtains an accuracy of 99.89%, a precision of 99.80%, and a recall of 99.72%, which places it in an excellent position for the categorization of leukemia. The proposed model outperformed several state-of-the-art Convolutional Neural Network (CNN) models in terms of performance. Consequently, this proposed model has the potential to save lives and effort. For a more comprehensive simulation of the entire methodology, a web application (Beta Version) has been developed in this study. This application is designed to determine the presence or absence of leukemia in individuals. The findings of this study hold significant potential for application in biomedical research, particularly in enhancing the accuracy of computer-aided leukemia detection.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    物联网(IoT)应用和资源极易受到洪水攻击,包括分布式拒绝服务(DDoS)攻击。这些攻击用大量网络数据包淹没了目标设备,使授权用户无法访问其资源。此类攻击可能包括攻击参考,攻击类型,子类别,主机信息,恶意脚本,等。这些细节有助于安全专业人员识别弱点,剪裁防御措施,并迅速应对可能的威胁,从而改善物联网设备的整体安全态势。由于其众多的网络特性,开发智能入侵检测系统(IDS)非常复杂。这项研究提出了一种改进的物联网安全IDS,它采用了多模式大数据表示和迁移学习。首先,会抓取数据包捕获(PCAP)文件以检索必要的攻击和字节。第二,基于Spark的大数据优化算法处理海量数据。第二,诸如word2vec之类的迁移学习方法检索基于语义的观察特征。第三,开发了一种将网络字节转换为图像的算法,和纹理特征通过配置基于注意力的残差网络(ResNet)来提取。最后,将训练好的文本和纹理特征组合起来,作为多模态特征对各种攻击进行分类。所提出的方法在三个广泛使用的基于物联网的数据集上进行了全面评估:CIC-IoT2022,CIC-IoT2023和Edge-IIoT。所提出的方法实现了优异的分类性能,准确率为98.2%。此外,我们提出了一个基于博弈论的过程来正式验证所提出的方法。
    Internet of Things (IoT) applications and resources are highly vulnerable to flood attacks, including Distributed Denial of Service (DDoS) attacks. These attacks overwhelm the targeted device with numerous network packets, making its resources inaccessible to authorized users. Such attacks may comprise attack references, attack types, sub-categories, host information, malicious scripts, etc. These details assist security professionals in identifying weaknesses, tailoring defense measures, and responding rapidly to possible threats, thereby improving the overall security posture of IoT devices. Developing an intelligent Intrusion Detection System (IDS) is highly complex due to its numerous network features. This study presents an improved IDS for IoT security that employs multimodal big data representation and transfer learning. First, the Packet Capture (PCAP) files are crawled to retrieve the necessary attacks and bytes. Second, Spark-based big data optimization algorithms handle huge volumes of data. Second, a transfer learning approach such as word2vec retrieves semantically-based observed features. Third, an algorithm is developed to convert network bytes into images, and texture features are extracted by configuring an attention-based Residual Network (ResNet). Finally, the trained text and texture features are combined and used as multimodal features to classify various attacks. The proposed method is thoroughly evaluated on three widely used IoT-based datasets: CIC-IoT 2022, CIC-IoT 2023, and Edge-IIoT. The proposed method achieves excellent classification performance, with an accuracy of 98.2%. In addition, we present a game theory-based process to validate the proposed approach formally.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    在材料的三维微观结构重建中建立精确的结构-性能联系和精确的相体积精度仍然具有挑战性,特别是有限的样本。本文提出了一种用于重建各种材料的3D微结构的优化方法。包括具有两相和三相的各向同性和各向异性类型,使用卷积占用网络和来自微观结构内层的点云。该方法强调精确的相位表示和与点云数据的兼容性。连接质量函数(QCF)重复循环中的一个阶段优化了卷积占用网络模型的权重,以最小化微观结构的统计属性与重建模型之间的误差。该模型成功地从初始2D系列图像重建3D表示。与筛选的泊松表面重建和局部隐式网格方法的比较证明了模型的有效性。所开发的模型证明适用于高质量的三维微结构重建,有助于结构-性能联系和有限元分析。
    Establishing accurate structure-property linkages and precise phase volume accuracy in 3D microstructure reconstruction of materials remains challenging, particularly with limited samples. This paper presents an optimized method for reconstructing 3D microstructures of various materials, including isotropic and anisotropic types with two and three phases, using convolutional occupancy networks and point clouds from inner layers of the microstructure. The method emphasizes precise phase representation and compatibility with point cloud data. A stage within the Quality of Connection Function (QCF) repetition loop optimizes the weights of the convolutional occupancy networks model to minimize error between the microstructure\'s statistical properties and the reconstructive model. This model successfully reconstructs 3D representations from initial 2D serial images. Comparisons with screened Poisson surface reconstruction and local implicit grid methods demonstrate the model\'s efficacy. The developed model proves suitable for high-quality 3D microstructure reconstruction, aiding in structure-property linkages and finite element analysis.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    蛋白质-肽相互作用(PPepIs)对于理解细胞功能至关重要。这可以促进新药的设计。作为形成PPepI的重要组成部分,蛋白-肽结合位点是理解PPepIs机制的基础.因此,准确识别蛋白-肽结合位点成为一项关键任务。传统的研究这些结合位点的实验方法费时费力,并且已经发明了一些计算工具来补充它。然而,由于需要配体信息,这些计算工具在通用性或准确性方面有限制,复杂的特征构造,或者他们对基于氨基酸残基的建模的依赖。为了解决这些计算算法的缺点,在这项工作中,我们描述了一个基于几何注意力的肽结合位点识别(GAPS)网络。所提出的模型利用几何特征工程来构建原子表示,并结合了多种注意力机制来更新相关的生物特征。此外,迁移学习策略是为了利用蛋白质-蛋白质结合位点信息来增强蛋白质-肽结合位点识别能力,考虑到蛋白质和肽之间的共同结构和生物学偏见。因此,GAPS在此任务中展示了最先进的性能和出色的鲁棒性。此外,我们的模型在几个扩展的实验中表现出卓越的性能,包括预测apo蛋白-肽,蛋白质环肽和AlphaFold预测的蛋白质肽结合位点。这些结果证实了GAPS模型是一个强大的,多才多艺,适用于多种结合位点预测的稳定方法。
    Protein-peptide interactions (PPepIs) are vital to understanding cellular functions, which can facilitate the design of novel drugs. As an essential component in forming a PPepI, protein-peptide binding sites are the basis for understanding the mechanisms involved in PPepIs. Therefore, accurately identifying protein-peptide binding sites becomes a critical task. The traditional experimental methods for researching these binding sites are labor-intensive and time-consuming, and some computational tools have been invented to supplement it. However, these computational tools have limitations in generality or accuracy due to the need for ligand information, complex feature construction, or their reliance on modeling based on amino acid residues. To deal with the drawbacks of these computational algorithms, we describe a geometric attention-based network for peptide binding site identification (GAPS) in this work. The proposed model utilizes geometric feature engineering to construct atom representations and incorporates multiple attention mechanisms to update relevant biological features. In addition, the transfer learning strategy is implemented for leveraging the protein-protein binding sites information to enhance the protein-peptide binding sites recognition capability, taking into account the common structure and biological bias between proteins and peptides. Consequently, GAPS demonstrates the state-of-the-art performance and excellent robustness in this task. Moreover, our model exhibits exceptional performance across several extended experiments including predicting the apo protein-peptide, protein-cyclic peptide and the AlphaFold-predicted protein-peptide binding sites. These results confirm that the GAPS model is a powerful, versatile, stable method suitable for diverse binding site predictions.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    肺癌是全球第二常见的癌症,也是导致癌症死亡的主要原因。低剂量计算机断层扫描(LDCT)是推荐的早期发现肺癌的影像学筛查工具。一种完全自动化的LDCT计算机辅助检测方法将极大地改善现有的临床工作流程。大多数现有的肺部检测方法都是为高剂量CT(HDCT)设计的,由于域移位和LDCT图像质量差,这些方法不能直接应用于LDCT。在这项工作中,我们描述了一种基于半自动化迁移学习的方法,用于使用LDCT早期检测肺结节.
    在这项工作中,我们开发了一种基于目标检测模型的算法,你只看一次(YOLO)来检测肺结节。YOLO模型首先是在CT上训练的,并且在使用医学到医学转移学习方法在LDCT上重新训练模型期间使用预训练权重作为初始权重。这项研究的数据集是来自一项筛选试验,该试验由从连续3年(T1,T2和T3)获得的50名活检证实的肺癌患者获得的LDCT组成。大约60名肺癌患者的HDCT来自一个公共数据集。使用包含15例患者病例(93个具有癌结节的切片)的固定测试集对开发的模型进行了评估,特异性,召回,和F1得分。评估指标每年按患者报告,并平均3年。为了进行比较分析,所提出的检测模型使用COCO数据集的预训练权重作为初始权重进行训练.采用α值为0.05的配对t检验和卡方检验进行统计学显著性检验。
    通过比较使用HDCT预训练权重与COCO预训练权重开发的拟议模型来报告结果。前一种方法与后一种方法在检测癌结节方面获得了0.982对0.93的精度,0.923与0.849的特异性在识别无癌结节的切片,召回率分别为0.87和0.886,F1评分为0.924和0.903。随着结节的进展,前一种方法的精密度为1,特异性为0.92,灵敏度为0.930.在对比研究中进行的统计分析导致精确度的p值为0.0054,特异性的p值为0.00034。
    在这项研究中,开发了一种半自动方法来检测LDCT中的肺结节,使用HDCT预训练权重作为初始权重并重新训练模型.Further,通过将上述方法中的HDCT预训练权重替换为COCO预训练权重来比较结果.所提出的方法可以在筛查程序中识别早期肺结节,减少由于LDCT误诊而导致的过度诊断和随访,在受影响的患者中开始治疗方案,降低死亡率。
    UNASSIGNED: Lung cancer is the second most common cancer and the leading cause of cancer death globally. Low dose computed tomography (LDCT) is the recommended imaging screening tool for the early detection of lung cancer. A fully automated computer-aided detection method for LDCT will greatly improve the existing clinical workflow. Most of the existing methods for lung detection are designed for high-dose CTs (HDCTs), and those methods cannot be directly applied to LDCTs due to domain shifts and inferior quality of LDCT images. In this work, we describe a semi-automated transfer learning-based approach for the early detection of lung nodules using LDCTs.
    UNASSIGNED: In this work, we developed an algorithm based on the object detection model, you only look once (YOLO) to detect lung nodules. The YOLO model was first trained on CTs, and the pre-trained weights were used as initial weights during the retraining of the model on LDCTs using a medical-to-medical transfer learning approach. The dataset for this study was from a screening trial consisting of LDCTs acquired from 50 biopsy-confirmed lung cancer patients obtained over 3 consecutive years (T1, T2, and T3). About 60 lung cancer patients\' HDCTs were obtained from a public dataset. The developed model was evaluated using a hold-out test set comprising 15 patient cases (93 slices with cancerous nodules) using precision, specificity, recall, and F1-score. The evaluation metrics were reported patient-wise on a per-year basis and averaged for 3 years. For comparative analysis, the proposed detection model was trained using pre-trained weights from the COCO dataset as the initial weights. A paired t-test and chi-squared test with an alpha value of 0.05 were used for statistical significance testing.
    UNASSIGNED: The results were reported by comparing the proposed model developed using HDCT pre-trained weights with COCO pre-trained weights. The former approach versus the latter approach obtained a precision of 0.982 versus 0.93 in detecting cancerous nodules, specificity of 0.923 versus 0.849 in identifying slices with no cancerous nodules, recall of 0.87 versus 0.886, and F1-score of 0.924 versus 0.903. As the nodule progressed, the former approach achieved a precision of 1, specificity of 0.92, and sensitivity of 0.930. The statistical analysis performed in the comparative study resulted in a p -value of 0.0054 for precision and a p -value of 0.00034 for specificity.
    UNASSIGNED: In this study, a semi-automated method was developed to detect lung nodules in LDCTs using HDCT pre-trained weights as the initial weights and retraining the model. Further, the results were compared by replacing HDCT pre-trained weights in the above approach with COCO pre-trained weights. The proposed method may identify early lung nodules during the screening program, reduce overdiagnosis and follow-ups due to misdiagnosis in LDCTs, start treatment options in the affected patients, and lower the mortality rate.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    目的:脑电图(EEG)被广泛认为是检测疲劳的有效方法。然而,EEG在现实世界场景中用于疲劳检测的实际应用通常具有挑战性,特别是在涉及未包含在训练数据集中的受试者的情况下,由于生物个体差异和嘈杂的标签。本研究旨在通过解决这些挑战,为跨学科疲劳检测开发一个有效的框架。
    方法:在本研究中,我们提出了一个新的框架,称为DP-MP,用于跨主题疲劳检测,它利用基于领域对抗神经网络(DANN)的原型表示与混合成对学习相结合。我们提出的DP-MP框架旨在通过在EEG信号中编码与疲劳相关的语义结构并探索跨个体的共享疲劳原型特征来减轻生物个体差异的影响。值得注意的是,据我们所知,这项工作是第一个概念化疲劳检测作为一个成对学习任务,从而有效地减少来自噪声标签的干扰。此外,我们在疲劳检测领域提出了混合成对学习(MixPa)方法,通过在样本之间引入更多样化和信息丰富的关系,拓宽了成对学习的优势。
    结果:在两个基准数据库上进行了跨学科实验,SEED-VIG和FTEF,实现最先进的性能,平均精度为88.14%和97.41%,分别。这些有希望的结果证明了我们模型的有效性和出色的泛化能力。
    结论:这是首次将基于EEG的疲劳检测概念化为成对学习任务,为这一领域提供了新的视角。此外,我们提出的DP-MP框架有效地解决了疲劳检测领域中生物个体差异和嘈杂标签的挑战,并展示了卓越的性能。我们的工作为未来的研究提供了宝贵的见解,促进脑机接口在现实场景中的疲劳检测应用。 .
    OBJECTIVE: Electroencephalography (EEG) is widely recognized as an effective method for detecting fatigue. However, practical applications of EEG for fatigue detection in real-world scenarios are often challenging, particularly in cases involving subjects not included in the training datasets, owing to bio-individual differences and noisy labels. This study aims to develop an effective framework for cross-subject fatigue detection by addressing these challenges.
    METHODS: In this study, we propose a novel framework, termed DP-MP, for cross-subject fatigue detection, which utilizes a Domain-Adversarial Neural Network (DANN)-based prototypical representation in conjunction with Mix-up pairwise learning. Our proposed DP-MP framework aims to mitigate the impact of bio-individual differences by encoding fatigue-related semantic structures within EEG signals and exploring shared fatigue prototype features across individuals. Notably, to the best of our knowledge, this work is the first to conceptualize fatigue detection as a pairwise learning task, thereby effectively reducing the interference from noisy labels. Furthermore, we propose the Mix-up pairwise learning (MixPa) approach in the field of fatigue detection, which broadens the advantages of pairwise learning by introducing more diverse and informative relationships among samples.
    RESULTS: Cross-subject experiments were conducted on two benchmark databases, SEED-VIG and FTEF, achieving state-of-the-art performance with average accuracies of 88.14% and 97.41%, respectively. These promising results demonstrate our model\'s effectiveness and excellent generalization capability.
    CONCLUSIONS: This is the first time EEG-based fatigue detection has been conceptualized as a pairwise learning task, offering a novel perspective to this field. Moreover, our proposed DP-MP framework effectively tackles the challenges of bio-individual differences and noisy labels in the fatigue detection field and demonstrates superior performance. Our work provides valuable insights for future research, promoting the application of brain-computer interfaces for fatigue detection in real-world scenarios. .
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    微调是迁移学习中的一项重要技术,在缺乏训练数据的任务中取得了显著的成功。然而,由于当源域和目标域之间的数据分布差异较大时,难以提取单源域微调的有效特征,我们提出了一种基于多源域的迁移学习框架,称为自适应多源域协作微调(AMCF)。AMCF利用多个源域模型进行协作微调,从而提高模型在目标任务中的特征提取能力。具体来说,AMCF采用自适应多源域层选择策略,为多个源域模型中的目标任务定制合适的层微调方案,旨在提取更有效的特征。此外,设计了一种新的多源域协同损失函数,便于各源域模型精确提取目标数据特征。同时,它致力于最小化各种源域模型之间的输出差异,增强了源域模型对目标数据的适应性。为了验证AMCF的有效性,它适用于迁移学习中常用的七个公共视觉分类数据集,并与最广泛使用的单源域微调方法进行了比较。实验结果表明,与现有的微调方法相比,我们的方法不仅提高了模型中特征提取的准确性,而且为目标任务提供了精确的层微调方案,从而显著提高微调性能。
    Fine-tuning is an important technique in transfer learning that has achieved significant success in tasks that lack training data. However, as it is difficult to extract effective features for single-source domain fine-tuning when the data distribution difference between the source and the target domain is large, we propose a transfer learning framework based on multi-source domain called adaptive multi-source domain collaborative fine-tuning (AMCF) to address this issue. AMCF utilizes multiple source domain models for collaborative fine-tuning, thereby improving the feature extraction capability of model in the target task. Specifically, AMCF employs an adaptive multi-source domain layer selection strategy to customize appropriate layer fine-tuning schemes for the target task among multiple source domain models, aiming to extract more efficient features. Furthermore, a novel multi-source domain collaborative loss function is designed to facilitate the precise extraction of target data features by each source domain model. Simultaneously, it works towards minimizing the output difference among various source domain models, thereby enhancing the adaptability of the source domain model to the target data. In order to validate the effectiveness of AMCF, it is applied to seven public visual classification datasets commonly used in transfer learning, and compared with the most widely used single-source domain fine-tuning methods. Experimental results demonstrate that, in comparison with the existing fine-tuning methods, our method not only enhances the accuracy of feature extraction in the model but also provides precise layer fine-tuning schemes for the target task, thereby significantly improving the fine-tuning performance.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    包含假脸的图像和视频是最常见的数字操纵类型。此类内容可能会通过传播虚假信息而导致负面后果。使用机器学习算法来生成假人脸图像使得区分真假内容变得很有挑战性。面部操作分为四个基本组:整个面部合成,面部身份操纵(deepfake),面部属性操纵和面部表情操纵。该研究利用轻量级卷积神经网络来检测使用整个人脸合成和生成对抗网络生成的假人脸图像。训练过程中使用的数据集包括FFHQ数据集中的70,000个真实图像和使用FFHQ数据集用StyleGAN2产生的70,000个假图像。80%的数据集用于训练,20%用于测试。最初,MobileNet,MobileNetV2、EfficientNetB0和NASNetMobile卷积神经网络被分别训练用于训练过程。在训练中,模型在ImageNet上进行了预训练,并与迁移学习一起重用.作为第一次训练的结果,EfficientNetB0算法达到了93.64%的最高精度。修改了EfficientNetB0算法,通过添加两个具有ReLU激活的密集层(256个神经元)来提高其准确率,两个dropout层,一个平坦层,一个具有ReLU激活功能的致密层(128个神经元),和用于具有两个节点的分类密集层的softmax激活函数。结果,EfficientNetB0算法实现了95.48%的过程准确率。最后,使用堆叠集成学习方法,将达到95.48%精度的模型用于一起训练MobileNet和MobileNetV2模型,的最高准确率为96.44%。
    Images and videos containing fake faces are the most common type of digital manipulation. Such content can lead to negative consequences by spreading false information. The use of machine learning algorithms to produce fake face images has made it challenging to distinguish between genuine and fake content. Face manipulations are categorized into four basic groups: entire face synthesis, face identity manipulation (deepfake), facial attribute manipulation and facial expression manipulation. The study utilized lightweight convolutional neural networks to detect fake face images generated by using entire face synthesis and generative adversarial networks. The dataset used in the training process includes 70,000 real images in the FFHQ dataset and 70,000 fake images produced with StyleGAN2 using the FFHQ dataset. 80% of the dataset was used for training and 20% for testing. Initially, the MobileNet, MobileNetV2, EfficientNetB0, and NASNetMobile convolutional neural networks were trained separately for the training process. In the training, the models were pre-trained on ImageNet and reused with transfer learning. As a result of the first trainings EfficientNetB0 algorithm reached the highest accuracy of 93.64%. The EfficientNetB0 algorithm was revised to increase its accuracy rate by adding two dense layers (256 neurons) with ReLU activation, two dropout layers, one flattening layer, one dense layer (128 neurons) with ReLU activation function, and a softmax activation function used for the classification dense layer with two nodes. As a result of this process accuracy rate of 95.48% was achieved with EfficientNetB0 algorithm. Finally, the model that achieved 95.48% accuracy was used to train MobileNet and MobileNetV2 models together using the stacking ensemble learning method, resulting in the highest accuracy rate of 96.44%.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

公众号