Model ensemble

  • 文章类型: Journal Article
    在需要大量车辆管理的行业中,对车辆损坏进行分类的自动化程序至关重要。尽管有大量的研究需求,由于公共数据集的稀缺性和构建数据集的复杂性,车辆损坏分类领域的挑战仍然存在。为了应对这些挑战,我们介绍了一个四分之三视图汽车伤害数据集(TQVCD数据集),强调标签的简单性,数据可访问性,和丰富的信息固有的四分之三的观点。TQVCD数据集通过车辆方向(前部或后部)和损坏类型来区分类别,同时保持四分之三的视图。我们使用五种流行的预训练深度学习架构——ResNet-50、DenseNet-160、EfficientNet-B0、MobileNet-V2和ViT——使用一套二元分类模型来评估性能。为了增强分类的鲁棒性,我们实现了一种模型集成方法,以有效地减轻单个模型的依赖\'偏差。此外,我们采访了二手车平台的三位专家,从工业角度验证了使用相应数据集建立车辆损伤分类模型的必要性。实证研究结果强调了数据集对车辆视角的全面覆盖,促进有效的数据收集和损坏分类,同时最大限度地减少劳动密集型标签工作。
    Automated procedures for classifying vehicle damage are critical in industries requiring extensive vehicle management. Despite substantial research demands, challenges in the field of vehicle damage classification persist due to the scarcity of public datasets and the complexity of constructing datasets. In response to these challenges, we introduce a Three-Quarter View Car Damage Dataset (TQVCD dataset), emphasizing simplicity in labeling, data accessibility, and rich information inherent in three-quarter views. The TQVCD dataset distinguishes class by vehicle orientation (front or rear) and type of damage while maintaining a three-quarter view. We evaluate performance using five prevalent pre-trained deep learning architectures-ResNet-50, DenseNet-160, EfficientNet-B0, MobileNet-V2, and ViT-employing a suite of binary classification models. To enhance classification robustness, we implement a model ensemble method to effectively mitigate individual model dependencies\' deviations. Additionally, we interview three experts from the used-car platform to validate the necessity of a vehicle damage classification model using the corresponding dataset from an industrial perspective. Empirical findings underscore the dataset\'s comprehensive coverage of vehicle perspectives, facilitating efficient data collection and damage classification while minimizing labor-intensive labeling efforts.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    由于人为活动造成的磷(P)过多,中国低地农村河流面临着严重的富营养化问题。然而,由于低地农村河流与周边地区的复杂相互作用,量化P动力学具有挑战性。为应对这一挑战,专门为低地农村河流设计了P动态模型(River-P)。该模型与环境流体动力学代码(EFDC)和低地Polder系统(PDP)的磷动力学模型相结合,以表征在低地农村河流疏dr影响下的P动力学。基于太湖流域代表性低地农村河流的两年(2020-2021)数据集,中国,对耦合模型进行了校准,并获得了总P(TP)浓度的模型性能(R2>0.59,RMSE<0.04mg/L)。我们在研究河流中的研究表明,(1)泥沙疏浚控制磷的有效性的时间尺度为300天,疏浚后P保留能力增加74.8kg/年,TP浓度降低23%。(2)疏浚显著降低沉积物磷释放量98%,虽然磷的再悬浮和沉降能力增加了16%和46%,分别。(3)沉积物-水界面(SWI)在河流内部磷的转移中起着至关重要的作用。由于重新暂停占TP进口的16%,结算占TP出口的47%。考虑到低地农村河流的巨大磷保留能力,带有大型植物的排水沟和池塘是提高磷保留能力的有希望的方法。我们的研究为当地环境部门提供了宝贵的见解,全面了解低地农村河流的磷动态。这样就可以评估P控制中沉积物疏浚的有效性并实施相应的P控制措施。
    China\'s lowland rural rivers are facing severe eutrophication problems due to excessive phosphorus (P) from anthropogenic activities. However, quantifying P dynamics in a lowland rural river is challenging due to its complex interaction with surrounding areas. A P dynamic model (River-P) was specifically designed for lowland rural rivers to address this challenge. This model was coupled with the Environmental Fluid Dynamics Code (EFDC) and the Phosphorus Dynamic Model for lowland Polder systems (PDP) to characterize P dynamics under the impact of dredging in a lowland rural river. Based on a two-year (2020-2021) dataset from a representative lowland rural river in the Lake Taihu Basin, China, the coupled model was calibrated and achieved a model performance (R2>0.59, RMSE<0.04 mg/L) for total P (TP) concentrations. Our research in the study river revealed that (1) the time scale for the effectiveness of sediment dredging for P control was ∼300 days, with an increase in P retention capacity by 74.8 kg/year and a decrease in TP concentrations of 23% after dredging. (2) Dredging significantly reduced P release from sediment by 98%, while increased P resuspension and settling capacities by 16% and 46%, respectively. (3) The sediment-water interface (SWI) plays a critical role in P transfer within the river, as resuspension accounts for 16% of TP imports, and settling accounts for 47% of TP exports. Given the large P retention capacity of lowland rural rivers, drainage ditches and ponds with macrophytes are promising approaches to enhance P retention capacity. Our study provides valuable insights for local environmental departments, allowing a comprehensive understanding of P dynamics in lowland rural rivers. This enable the evaluation of the efficacy of sediment dredging in P control and the implementation of corresponding P control measures.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    COVID-19的爆发以其相当迅速的传播震惊了整个世界,并挑战了不同的部门。限制其传播的最有效方法之一是对感染患者的早期和准确诊断。医学成像,如X射线和计算机断层扫描(CT),结合人工智能(AI)的潜力,在支持医务人员的诊断过程中起着至关重要的作用。因此,在这篇文章中,五种不同的深度学习模型(ResNet18,ResNet34,InceptionV3,InceptionResNetV2和DenseNet161)及其集合,使用多数投票,已被用于使用胸部X射线图像对COVID-19、肺炎和健康受试者进行分类。进行多标签分类以预测每位患者的多种病理,如果存在。首先,使用局部可解释性方法--遮挡,对每个网络的可解释性进行了彻底研究,显著性,输入X梯度,引导反向传播,积分梯度,和DeepLIFT-并使用全局技术-神经元激活谱。COVID-19分类模型的平均微F1得分在0.66至0.875之间,网络模型集合的平均微F1得分为0.89。定性结果表明,ResNets是最可解释的模型。这项研究证明了在做出有关最佳性能模型的决定之前,使用可解释性方法比较不同模型的重要性。
    The outbreak of COVID-19 has shocked the entire world with its fairly rapid spread, and has challenged different sectors. One of the most effective ways to limit its spread is the early and accurate diagnosing of infected patients. Medical imaging, such as X-ray and computed tomography (CT), combined with the potential of artificial intelligence (AI), plays an essential role in supporting medical personnel in the diagnosis process. Thus, in this article, five different deep learning models (ResNet18, ResNet34, InceptionV3, InceptionResNetV2, and DenseNet161) and their ensemble, using majority voting, have been used to classify COVID-19, pneumoniæ and healthy subjects using chest X-ray images. Multilabel classification was performed to predict multiple pathologies for each patient, if present. Firstly, the interpretability of each of the networks was thoroughly studied using local interpretability methods-occlusion, saliency, input X gradient, guided backpropagation, integrated gradients, and DeepLIFT-and using a global technique-neuron activation profiles. The mean micro F1 score of the models for COVID-19 classifications ranged from 0.66 to 0.875, and was 0.89 for the ensemble of the network models. The qualitative results showed that the ResNets were the most interpretable models. This research demonstrates the importance of using interpretability methods to compare different models before making a decision regarding the best performing model.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    多器官分割,识别和分离医学图像中的不同器官,是医学图象剖析中的一项根本任务。最近,深度学习的巨大成功促使其在多器官分割任务中得到广泛采用。然而,由于昂贵的劳动力成本和专业知识,多器官注释的可用性通常是有限的,因此在为基于深度学习的方法获得足够的训练数据方面提出了挑战。在本文中,我们的目标是通过结合现成的单器官分割模型在目标数据集上开发多器官分割模型来解决这个问题,这有助于摆脱多器官分割对注释数据的依赖。为此,我们提出了一种新颖的双阶段方法,该方法由模型适应阶段和模型集成阶段组成。第一阶段增强了每个现成分割模型在目标域上的泛化,而第二阶段从多个适应的单器官分割模型中提取和整合知识。在四个腹部数据集上的大量实验表明,我们提出的方法可以有效地利用现成的单器官分割模型来获得高精度的多器官分割模型。
    Multi-organ segmentation, which identifies and separates different organs in medical images, is a fundamental task in medical image analysis. Recently, the immense success of deep learning motivated its wide adoption in multi-organ segmentation tasks. However, due to expensive labor costs and expertise, the availability of multi-organ annotations is usually limited and hence poses a challenge in obtaining sufficient training data for deep learning-based methods. In this paper, we aim to address this issue by combining off-the-shelf single-organ segmentation models to develop a multi-organ segmentation model on the target dataset, which helps get rid of the dependence on annotated data for multi-organ segmentation. To this end, we propose a novel dual-stage method that consists of a Model Adaptation stage and a Model Ensemble stage. The first stage enhances the generalization of each off-the-shelf segmentation model on the target domain, while the second stage distills and integrates knowledge from multiple adapted single-organ segmentation models. Extensive experiments on four abdomen datasets demonstrate that our proposed method can effectively leverage off-the-shelf single-organ segmentation models to obtain a tailored model for multi-organ segmentation with high accuracy.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 识别低流行疾病,或罕见疾病,早期是医学领域疾病治疗的关键。深度学习技术现在为此目的提供了有前途的工具。然而,由于严重的类不平衡问题,罕见疾病的低患病率困扰着深度网络在疾病识别中的正确应用。在过去的几十年里,已经研究了一些平衡方法来处理数据不平衡问题。坏消息是,事实证明,这些方法都不能保证比其他方法优越的性能。这种性能差异导致需要使用全面的软件工具制定系统的管道,以增强罕见疾病识别中的深度学习应用。我们回顾了现有的平衡方案,并总结了一个系统的深层集成管道,并使用一个名为RDDL的构建工具来处理数据不平衡问题。通过两个真实的案例研究,我们表明,通过减少模型训练过程中的数据不平衡问题,可以使用这种系统的RDDL管道工具来增强罕见疾病的识别。RDDL管道工具可在https://github.com/cobisLab/RDDL/上获得。
    Identifying lowly prevalent diseases, or rare diseases, in their early stages is key to disease treatment in the medical field. Deep learning techniques now provide promising tools for this purpose. Nevertheless, the low prevalence of rare diseases entangles the proper application of deep networks for disease identification due to the severe class-imbalance issue. In the past decades, some balancing methods have been studied to handle the data-imbalance issue. The bad news is that it is verified that none of these methods guarantees superior performance to others. This performance variation causes the need to formulate a systematic pipeline with a comprehensive software tool for enhancing deep-learning applications in rare disease identification. We reviewed the existing balancing schemes and summarized a systematic deep ensemble pipeline with a constructed tool called RDDL for handling the data imbalance issue. Through two real case studies, we showed that rare disease identification could be boosted with this systematic RDDL pipeline tool by lessening the data imbalance problem during model training. The RDDL pipeline tool is available at https://github.com/cobisLab/RDDL/.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    基于Transformer的语言模型已成功用于解决大量与文本相关的任务。DNA甲基化是一种重要的表观遗传机制,它的分析为基因调控和生物标志物鉴定提供了有价值的见解。已经提出了几种基于深度学习的方法来识别DNA甲基化,每个人都寻求在计算工作量和准确性之间取得平衡。这里,我们介绍MuLan-甲基,预测DNA甲基化位点的深度学习框架,它基于5种流行的基于变压器的语言模型。该框架确定了3种不同类型的DNA甲基化的甲基化位点:N6-腺嘌呤,N4-胞嘧啶,和5-羟甲基胞嘧啶。所采用的语言模型中的每一个都使用“预训练和微调”范例来适应任务。使用自监督学习对DNA片段和分类谱系的自定义语料库进行预训练。微调旨在预测每种类型的DNA甲基化状态。5个模型用于共同预测DNA甲基化状态。我们在基准数据集上报告了MuLan-Methyl的出色性能。此外,我们认为该模型捕获了与甲基化相关的不同物种之间的特征差异。这项工作表明,语言模型可以成功地适应生物序列分析中的应用程序,并且联合使用不同的语言模型可以提高模型性能。木兰甲基是开源的,我们提供了一个实现该方法的Web服务器。
    Transformer-based language models are successfully used to address massive text-related tasks. DNA methylation is an important epigenetic mechanism, and its analysis provides valuable insights into gene regulation and biomarker identification. Several deep learning-based methods have been proposed to identify DNA methylation, and each seeks to strike a balance between computational effort and accuracy. Here, we introduce MuLan-Methyl, a deep learning framework for predicting DNA methylation sites, which is based on 5 popular transformer-based language models. The framework identifies methylation sites for 3 different types of DNA methylation: N6-adenine, N4-cytosine, and 5-hydroxymethylcytosine. Each of the employed language models is adapted to the task using the \"pretrain and fine-tune\" paradigm. Pretraining is performed on a custom corpus of DNA fragments and taxonomy lineages using self-supervised learning. Fine-tuning aims at predicting the DNA methylation status of each type. The 5 models are used to collectively predict the DNA methylation status. We report excellent performance of MuLan-Methyl on a benchmark dataset. Moreover, we argue that the model captures characteristic differences between different species that are relevant for methylation. This work demonstrates that language models can be successfully adapted to applications in biological sequence analysis and that joint utilization of different language models improves model performance. Mulan-Methyl is open source, and we provide a web server that implements the approach.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    本文提出了通过组合模型和不同模型的变体来进行深度学习对象检测模型的集成策略,以增强脑MRI中解剖和病理对象的检测性能。在这项研究中,在新的GaziBrains2020数据集的帮助下,5个不同的解剖部位和一个病理部位,可以在脑MRI中观察到,例如感兴趣的区域,眼睛,视神经,侧脑室,第三脑室,还有一个完整的肿瘤.首先,对九种最先进的对象检测模型进行了全面的基准测试,以确定模型在检测解剖和病理部位方面的能力。然后,应用了9个对象检测器的四种不同的集成策略,以使用边界框融合技术来提高检测性能。单个模型变体的集合在平均平均精度(mAP)方面将解剖和病理对象检测性能提高了多达10%。此外,考虑解剖部位的基于类的平均精度(AP)值,实现了高达18%的AP改善。同样,最佳不同模型的集成策略优于最佳个体模型3.3%的mAP。此外,而高达7%的FAUC更好,这是TPR下的区域FPPI曲线,是在GaziBrains2020数据集上实现的,在BraTS2020数据集上获得的FAUC评分提高了2%.所提出的集合策略被发现在找到具有少量解剖对象的解剖和病理部分方面更有效,比如视神经和第三脑室,产生更高的TPR值,特别是在低FPPI值,与最好的个人方法相比。
    This paper proposes ensemble strategies for the deep learning object detection models carried out by combining the variants of a model and different models to enhance the anatomical and pathological object detection performance in brain MRI. In this study, with the help of the novel Gazi Brains 2020 dataset, five different anatomical parts and one pathological part that can be observed in brain MRI were identified, such as the region of interest, eye, optic nerves, lateral ventricles, third ventricle, and a whole tumor. Firstly, comprehensive benchmarking of the nine state-of-the-art object detection models was carried out to determine the capabilities of the models in detecting the anatomical and pathological parts. Then, four different ensemble strategies for nine object detectors were applied to boost the detection performance using the bounding box fusion technique. The ensemble of individual model variants increased the anatomical and pathological object detection performance by up to 10% in terms of the mean average precision (mAP). In addition, considering the class-based average precision (AP) value of the anatomical parts, an up to 18% AP improvement was achieved. Similarly, the ensemble strategy of the best different models outperformed the best individual model by 3.3% mAP. Additionally, while an up to 7% better FAUC, which is the area under the TPR vs. FPPI curve, was achieved on the Gazi Brains 2020 dataset, a 2% better FAUC score was obtained on the BraTS 2020 dataset. The proposed ensemble strategies were found to be much more efficient in finding the anatomical and pathological parts with a small number of anatomic objects, such as the optic nerve and third ventricle, and producing higher TPR values, especially at low FPPI values, compared to the best individual methods.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    该研究预测了上印度河盆地(UIB)的气候,覆盖印度的地理区域,巴基斯坦,阿富汗,和中国,在两个代表性集中途径(RCP)下,viz.,到21世纪后期,RCP4.5和RCP8.5使用最适合的气候模型,根据八个气象站的气候观测结果进行了验证。GFDLCM3在模拟UIB气候方面的表现优于其他五个评估的气候模型。通过Aerts和Droogers统计降尺度方法,模型偏差显著降低,总体预测显示,整个由Jhelum组成的UIB的温度显着增加,降水略有增加,Chenab,和印度河次盆地。根据RCP4.5和RCP8.5,到21世纪后期,Jhelum的温度和降水预计将分别增加3°C和5.2°C以及0.8%和3.4%。在这两种情况下,到21世纪后期,Chenab的温度和降水预计将分别增加3.5°C和4.8°C以及8%和8.2%。在RCP4.5和RCP8.5情景下,到21世纪后期,印度河的温度和降水预计将分别增加4.8°C和6.5°C以及2.6%和8.7%。二十一世纪后期预计的气候将对各种生态系统服务和产品产生重大影响,灌溉和社会水文制度,和各种依赖的生计。因此,希望高分辨率气候预测将有助于影响评估研究,为UIB的气候行动决策提供信息。
    The study projects climate over the Upper Indus Basin (UIB), covering geographic areas in India, Pakistan, Afghanistan, and China, under the two Representative Concentration Pathways (RCPs), viz., RCP4.5 and RCP8.5 by the late twenty-first century using the best-fit climate model validated against the climate observations from eight meteorological stations. GFDL CM3 performed better than the other five evaluated climate models in simulating the climate of the UIB. The model bias was significantly reduced by the Aerts and Droogers statistical downscaling method, and the projections overall revealed a significant increase in temperature and a slight increase in precipitation across the UIB comprising of Jhelum, Chenab, and Indus sub-basins. According to RCP4.5 and RCP8.5, the temperature and precipitation in the Jhelum are projected to increase by 3 °C and 5.2 °C and 0.8% and 3.4% respectively by the late twenty-first century. The temperature and precipitation in the Chenab are projected to increase by 3.5 °C and 4.8 °C and  8% and 8.2% respectively by the late twenty-first century under the two scenarios. The temperature and precipitation in the Indus are projected to increase by 4.8 °C and 6.5 °C and 2.6% and 8.7% respectively by the late twenty-first century under RCP4.5 and RCP8.5 scenarios. The late twenty-first century projected climate would have significant impacts on various ecosystem services and products, irrigation and socio-hydrological regimes, and various dependent livelihoods. It is therefore hoped that the high-resolution climate projections would be useful for impact assessment studies to inform policymaking for climate action in the UIB.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    背景:获得口腔医疗保健在全球范围内并不统一,特别是在资源有限的农村地区,这限制了自动诊断和先进的远程牙科应用的潜力。通过照相通信使用数字龋齿检测和进展监测,受到在此类设置中难以标准化的多个变量的影响。这项研究的目的是开发一种新颖且具有成本效益的虚拟计算机视觉AI系统,以合理的临床准确性从非标准化照片中预测牙洞。
    方法:从233个去识别的牙齿标本中获得了一组1703个增强图像。使用消费者智能手机获取图像,没有任何标准化的设备应用。这项研究利用了最先进的合奏建模,测试时间增加,和迁移学习过程。“你只看一次”算法(YOLO)衍生物,v5s,v5m,v5l,和v5x,进行了独立评估,增强了最佳结果的集合,以及使用ResNet50、ResNet101、VGG16、AlexNet学习的传输,和DenseNet。使用精确度评估结果,召回,和平均平均精度(MAP)。
    结果:YOLO模型集合的平均精度(mAP)为0.732,准确率为0.789,召回率为0.701。当转移到VGG16时,最终模型的诊断准确率为86.96%,准确率为0.89,召回率为0.88。这超过了所有其他从免费的非标准化智能手机照片中进行物体检测的基本方法。
    结论:虚拟计算机视觉AI系统,混合模型集合,测试时间增加,转移了深度学习过程,开发用于从具有合理临床准确性的非标准化照片中预测牙洞。这种模式可以改善农村地区在资源有限的情况下获得口腔保健的机会,并有可能帮助自动诊断和先进的远程牙科应用。
    Access to oral healthcare is not uniform globally, particularly in rural areas with limited resources, which limits the potential of automated diagnostics and advanced tele-dentistry applications. The use of digital caries detection and progression monitoring through photographic communication, is influenced by multiple variables that are difficult to standardize in such settings. The objective of this study was to develop a novel and cost-effective virtual computer vision AI system to predict dental cavitations from non-standardised photographs with reasonable clinical accuracy.
    A set of 1703 augmented images was obtained from 233 de-identified teeth specimens. Images were acquired using a consumer smartphone, without any standardised apparatus applied. The study utilised state-of-the-art ensemble modeling, test-time augmentation, and transfer learning processes. The \"you only look once\" algorithm (YOLO) derivatives, v5s, v5m, v5l, and v5x, were independently evaluated, and an ensemble of the best results was augmented, and transfer learned with ResNet50, ResNet101, VGG16, AlexNet, and DenseNet. The outcomes were evaluated using precision, recall, and mean average precision (mAP).
    The YOLO model ensemble achieved a mean average precision (mAP) of 0.732, an accuracy of 0.789, and a recall of 0.701. When transferred to VGG16, the final model demonstrated a diagnostic accuracy of 86.96%, precision of 0.89, and recall of 0.88. This surpassed all other base methods of object detection from free-hand non-standardised smartphone photographs.
    A virtual computer vision AI system, blending a model ensemble, test-time augmentation, and transferred deep learning processes, was developed to predict dental cavitations from non-standardised photographs with reasonable clinical accuracy. This model can improve access to oral healthcare in rural areas with limited resources, and has the potential to aid in automated diagnostics and advanced tele-dentistry applications.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    抗癌肽(ACP)是已被证明具有抗癌活性的肽类型。使用ACP预防癌症可能是传统癌症治疗的可行替代方案,因为它们更安全且显示出更高的选择性。由于ACP鉴定是高度实验室受限的,昂贵而冗长,在这项研究中,提出了一种根据序列信息预测ACP的计算方法。该过程包括肽序列的输入,根据具有位置信息和手工特征的顺序编码进行特征提取,最后是特征选择。整个模型由两个模块组成,包括深度学习和机器学习算法。深度学习模块包含两个通道:双向长短期记忆(BiLSTM)和卷积神经网络(CNN)。在机器学习模块中使用光梯度提升机(LightGBM)。最后,本研究对三个模型的分类结果进行了投票,得出了模型集成层的三个路径。这项研究利用一种新颖的方法提供了对ACP预测的见解,并提出了有希望的性能。与以前的研究相比,它使用了基准数据集进行进一步的探索和改进。我们的最终模型的准确性为0.7895,灵敏度为0.8153,特异性为0.7676,与所有指标的最新研究相比,它至少增加了2%。因此,本文提出了一种新的方法,可以更有效地预测ACP。这些工作和源代码可通过https://github.com/khanhlee/acp-ope/向研究人员和开发人员社区提供。
    Anticancer peptides (ACPs) are the types of peptides that have been demonstrated to have anticancer activities. Using ACPs to prevent cancer could be a viable alternative to conventional cancer treatments because they are safer and display higher selectivity. Due to ACP identification being highly lab-limited, expensive and lengthy, a computational method is proposed to predict ACPs from sequence information in this study. The process includes the input of the peptide sequences, feature extraction in terms of ordinal encoding with positional information and handcrafted features, and finally feature selection. The whole model comprises of two modules, including deep learning and machine learning algorithms. The deep learning module contained two channels: bidirectional long short-term memory (BiLSTM) and convolutional neural network (CNN). Light Gradient Boosting Machine (LightGBM) was used in the machine learning module. Finally, this study voted the three models\' classification results for the three paths resulting in the model ensemble layer. This study provides insights into ACP prediction utilizing a novel method and presented a promising performance. It used a benchmark dataset for further exploration and improvement compared with previous studies. Our final model has an accuracy of 0.7895, sensitivity of 0.8153 and specificity of 0.7676, and it was increased by at least 2% compared with the state-of-the-art studies in all metrics. Hence, this paper presents a novel method that can potentially predict ACPs more effectively and efficiently. The work and source codes are made available to the community of researchers and developers at https://github.com/khanhlee/acp-ope/.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

公众号