Class imbalance

  • 文章类型: Journal Article
    识别属于不同类别的恒星对于建立恒星演化的不同阶段和路径的统计样本至关重要。在调查涉及数十亿颗恒星的时代,识别这些类的自动化方法变得必要。
    根据其发射的光谱来识别许多类别的恒星。在本文中,我们使用多类多标签机器学习(ML)方法XGBoost和PySSED光谱能量分布拟合算法的组合将恒星分为9个不同的类别,基于他们的光度数据.在SIMBAD数据库的子集上训练分类器。特别的挑战是底层数据的非常高的稀疏性(大部分缺失值)以及高类不平衡。我们讨论可用的不同变量,例如一方面的光度测量,以及间接预测因子,例如银河位置。
    我们显示了排除某些变量时的性能差异,并讨论应在哪些上下文中使用哪些变量。最后,我们表明,增加特定类型的恒星的样本数量显着增加该特定类型的模型的性能,而对其他类型几乎没有影响。主分类器的准确率为0.7,宏F1评分为0.61。
    尽管分类器的当前精度还不够高,无法可靠地用于恒星分类,这项工作是使用ML基于测光对恒星进行分类的可行性的初步证明。
    天文学处于“大数据”制度的最前沿,随着望远镜收集越来越大量的数据。天文学家用来分析和从这些数据中得出结论的工具需要能够跟上,机器学习提供了许多解决方案。能够按类型对不同的天文物体进行分类有助于解开天体物理学,使它们独一无二,为宇宙如何运作提供新的见解。这里,我们介绍了如何使用机器学习来对不同种类的恒星进行分类,为了扩大天空的大型数据库。这将使天文学家能够更轻松地提取他们进行科学分析所需的数据。
    UNASSIGNED: Identifying stars belonging to different classes is vital in order to build up statistical samples of different phases and pathways of stellar evolution. In the era of surveys covering billions of stars, an automated method of identifying these classes becomes necessary.
    UNASSIGNED: Many classes of stars are identified based on their emitted spectra. In this paper, we use a combination of the multi-class multi-label Machine Learning (ML) method XGBoost and the PySSED spectral-energy-distribution fitting algorithm to classify stars into nine different classes, based on their photometric data. The classifier is trained on subsets of the SIMBAD database. Particular challenges are the very high sparsity (large fraction of missing values) of the underlying data as well as the high class imbalance. We discuss the different variables available, such as photometric measurements on the one hand, and indirect predictors such as Galactic position on the other hand.
    UNASSIGNED: We show the difference in performance when excluding certain variables, and discuss in which contexts which of the variables should be used. Finally, we show that increasing the number of samples of a particular type of star significantly increases the performance of the model for that particular type, while having little to no impact on other types. The accuracy of the main classifier is ∼0.7 with a macro F1 score of 0.61.
    UNASSIGNED: While the current accuracy of the classifier is not high enough to be reliably used in stellar classification, this work is an initial proof of feasibility for using ML to classify stars based on photometry.
    Astronomy is at the forefront of the ‘Big Data’ regime, with telescopes collecting increasingly large volumes of data. The tools astronomers use to analyse and draw conclusions from these data need to be able to keep up, with machine learning providing many of the solutions. Being able to classify different astronomical objects by type helps to disentangle the astrophysics making them unique, offering new insights into how the Universe works. Here, we present how machine learning can be used to classify different kinds of stars, in order to augment large databases of the sky. This will allow astronomers to more easily extract the data they need to perform their scientific analyses.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    2型糖尿病(T2D)是一种普遍存在的终身健康状况。据预测,到2040年,将有超过5亿成年人被诊断为T2D。T2D可以在任何年龄发展,如果进展,它可能会导致严重的合并症。与T2D相关的最关键的合并症之一是心肌梗死(MI),被称为心脏病发作。MI是危及生命的紧急医疗事件,重要的是要预测它并及时干预。使用机器学习(ML)进行临床预测的步伐正在加快,但是预测模型中的类不平衡是建立可信赖的技术部署的关键挑战。这可能会导致ML模型中的偏差和过度拟合,它可能会导致对ML输出的误导性解释。在我们的研究中,我们展示了如何系统地使用类不平衡处理(CIH)技术可以提高ML模型的性能。我们使用了ConnectedBradford数据集,由超过一百万的真实健康记录组成.三种常用的CIH技术,过采样,欠采样,和类别加权(CW)已用于朴素贝叶斯(NB),神经网络(NN),随机森林(RF),支持向量机(SVM)和合奏模型。我们报告说,在其他技术中,CW的表现优于其他技术,其最高的准确性和F1值分别为0.9948和0.9556。使用真实世界的医疗保健数据将最适当的CIH技术应用于ML模型提供了有希望的结果,有助于降低T2D患者的MI风险。
    Type 2 Diabetes (T2D) is a prevalent lifelong health condition. It is predicted that over 500 million adults will be diagnosed with T2D by 2040. T2D can develop at any age, and if it progresses, it may cause serious comorbidities. One of the most critical T2D-related comorbidities is Myocardial Infarction (MI), known as heart attack. MI is a life-threatening medical emergency, and it is important to predict it and intervene in a timely manner. The use of Machine Learning (ML) for clinical prediction is gaining pace, but the class imbalance in predictive models is a key challenge for establishing a trustworthy deployment of the technology. This may lead to bias and overfitting in the ML models, and it may cause misleading interpretations of the ML outputs. In our study, we showed how systematic use of Class Imbalance Handling (CIH) techniques may improve the performance of the ML models. We used the Connected Bradford dataset, consisting of over one million real-world health records. Three commonly used CIH techniques, Oversampling, Undersampling, and Class Weighting (CW) have been used for Naive Bayes (NB), Neural Network (NN), Random Forest (RF), Support Vector Machine (SVM), and Ensemble models. We report that CW overperforms among the other techniques with the highest Accuracy and F1 values of 0.9948 and 0.9556, respectively. Applying the most appropriate CIH techniques for the ML models using real-world healthcare data provides promising results for helping to reduce the risk of MI in patients with T2D.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    在临床实践中,肺静脉的解剖分类在房颤射频消融手术的术前评估中起着至关重要的作用。肺静脉解剖结构的准确分类有助于医生选择合适的标测电极,避免引起肺动脉高压。由于肺静脉的解剖分类多种多样,以及数据分布的不平衡,深度学习模型在提取深度特征时往往表现出较差的表达能力,导致误判,影响分类精度。因此,为了解决左心房肺静脉分类不平衡的问题,本文提出了一种融合多尺度特征增强注意力和双特征提取分类器的网络,叫做DECNet。多尺度特征增强注意力利用多尺度信息引导深层特征的强化,生成通道权重和空间权重,增强深层特征的表达能力。双特征提取分类器为每个类别分配固定数量的通道,平等地评估所有类别,从而缓解了数据失衡导致的学习偏差和过拟合。通过将两者结合起来,增强了深层特征的表达能力,实现对左心房肺静脉形态的准确分类,为后续临床治疗提供支持。所提出的方法是在辽宁省人民医院提供的数据集和公开的DermaMNIST数据集上进行评估的,平均准确率为78.81%和83.44%,分别,证明了所提出方法的有效性。
    In clinical practice, the anatomical classification of pulmonary veins plays a crucial role in the preoperative assessment of atrial fibrillation radiofrequency ablation surgery. Accurate classification of pulmonary vein anatomy assists physicians in selecting appropriate mapping electrodes and avoids causing pulmonary arterial hypertension. Due to the diverse and subtly different anatomical classifications of pulmonary veins, as well as the imbalance in data distribution, deep learning models often exhibit poor expression capability in extracting deep features, leading to misjudgments and affecting classification accuracy. Therefore, in order to solve the problem of unbalanced classification of left atrial pulmonary veins, this paper proposes a network integrating multi-scale feature-enhanced attention and dual-feature extraction classifiers, called DECNet. The multi-scale feature-enhanced attention utilizes multi-scale information to guide the reinforcement of deep features, generating channel weights and spatial weights to enhance the expression capability of deep features. The dual-feature extraction classifier assigns a fixed number of channels to each category, equally evaluating all categories, thus alleviating the learning bias and overfitting caused by data imbalance. By combining the two, the expression capability of deep features is strengthened, achieving accurate classification of left atrial pulmonary vein morphology and providing support for subsequent clinical treatment. The proposed method is evaluated on datasets provided by the People\'s Hospital of Liaoning Province and the publicly available DermaMNIST dataset, achieving average accuracies of 78.81% and 83.44%, respectively, demonstrating the effectiveness of the proposed approach.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    在乌干达,缺乏用于构建机器学习模型来预测口蹄疫暴发的统一数据集阻碍了防备。尽管机器学习模型在固定条件下对口蹄疫暴发表现出出色的预测性能,它们在非平稳环境中容易受到性能下降的影响。降雨和温度是影响这些疫情爆发的关键因素,以及它们由于气候变化而产生的变异性会显著影响预测性能。这项研究创建了一个统一的口蹄疫数据集,通过整合不同的来源和预处理数据使用平均归因,重复删除,可视化,和合并技术。要评估性能下降,使用包括准确性在内的指标对七个机器学习模型进行了训练和评估,接收器工作特性曲线下的面积,召回,精度和F1分数。数据集显示了严重的类不平衡,非爆发多于爆发,需要数据增强方法。降雨和温度的变化影响预测性能,导致明显的退化。具有边界SMOTE的随机森林是静止环境中表现最好的模型,达到92%的准确率,接收器工作特性曲线下的0.97面积,0.94召回,0.90精度,和0.92F1分数。然而,在不同的分布下,所有模型都表现出显著的性能下降,随机森林准确率下降到46%,受试者工作特征曲线下面积为0.58,召回率为0.03,精度为0.24,F1评分为0.06。这项研究强调了为乌干达创建统一的口蹄疫数据集,并揭示了七个机器学习模型在不同分布下的显着性能下降。这些发现强调了需要新的方法来解决分布变异性对预测性能的影响。
    In Uganda, the absence of a unified dataset for constructing machine learning models to predict Foot and Mouth Disease outbreaks hinders preparedness. Although machine learning models exhibit excellent predictive performance for Foot and Mouth Disease outbreaks under stationary conditions, they are susceptible to performance degradation in non-stationary environments. Rainfall and temperature are key factors influencing these outbreaks, and their variability due to climate change can significantly impact predictive performance. This study created a unified Foot and Mouth Disease dataset by integrating disparate sources and pre-processing data using mean imputation, duplicate removal, visualization, and merging techniques. To evaluate performance degradation, seven machine learning models were trained and assessed using metrics including accuracy, area under the receiver operating characteristic curve, recall, precision and F1-score. The dataset showed a significant class imbalance with more non-outbreaks than outbreaks, requiring data augmentation methods. Variability in rainfall and temperature impacted predictive performance, causing notable degradation. Random Forest with borderline SMOTE was the top-performing model in a stationary environment, achieving 92% accuracy, 0.97 area under the receiver operating characteristic curve, 0.94 recall, 0.90 precision, and 0.92 F1-score. However, under varying distributions, all models exhibited significant performance degradation, with random forest accuracy dropping to 46%, area under the receiver operating characteristic curve to 0.58, recall to 0.03, precision to 0.24, and F1-score to 0.06. This study underscores the creation of a unified Foot and Mouth Disease dataset for Uganda and reveals significant performance degradation in seven machine learning models under varying distributions. These findings highlight the need for new methods to address the impact of distribution variability on predictive performance.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    背景:这项研究的目的是通过使用与机器参数和类平衡技术相关的输入特征来提高深度学习(DL)模型在预测和分类IMRT伽马通过率(GPR)方面的性能。
    方法:回顾性收集来自204例鼻咽癌患者IMRT计划的2348个字段,形成一个数据集。输入特征映射,包括注量,叶间隙,两岸的叶片速度,和相应的错误,是从动态日志文件构造的。SHAP框架用于计算每个特征对递归特征消除的模型输出的影响。一系列基于UNet++的模型在获得的八个特征集上进行了训练,使用三种微调方法,包括标准均方误差(MSE)损失,一种重新采样技术,和拟议的加权MSE损失(WMSE)。平均绝对误差的差异,接收器工作特性曲线下面积(AUC),灵敏度,比较了不同模型之间的特异性。
    结果:使用包括叶片速度和叶片间隙特征的特征集训练的模型比其他模型更准确地预测了失败领域的GPR(F(7,147)=5.378,p<0.001)。在三种微调方法中,WMSE损失在预测失败场的GPR方面具有最高的准确性(F(2,42)=14.149,p<0.001),而在预测通过场的GPR时观察到相反的趋势(F(2,730)=9.907,p<0.001)。与其他模型相比,WMSE_FS5模型实现了优异的AUC(0.92)和更平衡的灵敏度(0.77)和特异性(0.89)。
    结论:机器参数可以为DL中的GPR预测提供有区别的输入特征。新颖的加权损失函数证明了在通过和失败的字段之间平衡预测和分类准确性的能力。所提出的方法能够提高DL模型在预测和分类GPR方面的性能,并且可以潜在地集成到计划优化过程中,以生成更高的可交付性计划。
    背景:该临床试验于3月26日在中国临床试验注册中心注册,2020年(注册号:ChiCTR2000031276)。https://clinicaltrials.gov/ct2/show/ChiCTR2000031276.
    BACKGROUND: The purpose of this study was to improve the deep learning (DL) model performance in predicting and classifying IMRT gamma passing rate (GPR) by using input features related to machine parameters and a class balancing technique.
    METHODS: A total of 2348 fields from 204 IMRT plans for patients with nasopharyngeal carcinoma were retrospectively collected to form a dataset. Input feature maps, including fluence, leaf gap, leaf speed of both banks, and corresponding errors, were constructed from the dynamic log files. The SHAP framework was employed to compute the impact of each feature on the model output for recursive feature elimination. A series of UNet++ based models were trained on the obtained eight feature sets with three fine-tuning methods including the standard mean squared error (MSE) loss, a re-sampling technique, and a proposed weighted MSE loss (WMSE). Differences in mean absolute error, area under the receiver operating characteristic curve (AUC), sensitivity, and specificity were compared between the different models.
    RESULTS: The models trained with feature sets including leaf speed and leaf gap features predicted GPR for failed fields more accurately than the other models (F(7, 147) = 5.378, p < 0.001). The WMSE loss had the highest accuracy in predicting GPR for failed fields among the three fine-tuning methods (F(2, 42) = 14.149, p < 0.001), while an opposite trend was observed in predicting GPR for passed fields (F(2, 730) = 9.907, p < 0.001). The WMSE_FS5 model achieved a superior AUC (0.92) and more balanced sensitivity (0.77) and specificity (0.89) compared to the other models.
    CONCLUSIONS: Machine parameters can provide discriminative input features for GPR prediction in DL. The novel weighted loss function demonstrates the ability to balance the prediction and classification accuracy between the passed and failed fields. The proposed approach is able to improve the DL model performance in predicting and classifying GPR, and can potentially be integrated into the plan optimization process to generate higher deliverability plans.
    BACKGROUND: This clinical trial was registered in the Chinese Clinical Trial Registry on March 26th, 2020 (registration number: ChiCTR2000031276). https://clinicaltrials.gov/ct2/show/ChiCTR2000031276.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    脑胶质瘤的准确预测和分级在评估脑肿瘤的进展中起着至关重要的作用。评估总体预后,和治疗计划。除了神经成像技术,确定可以指导诊断的分子生物标志物,对治疗反应的预测和预测引起了研究人员对它们与机器学习和深度学习模型一起使用的兴趣。该领域的大部分研究都是以模型为中心,这意味着它是基于找到性能更好的算法。然而,在实践中,提高数据质量可以产生更好的模型。这项研究调查了一种以数据为中心的机器学习方法,以确定它们在预测神经胶质瘤等级方面的潜在益处。我们报告了六个性能指标,以提供模型性能的完整图景。实验结果表明,标准化和过度调整少数类增加了四个流行的机器学习模型和两个分类器集成应用于由临床因素和分子生物标记组成的低不平衡数据集的预测性能。实验还表明,两个分类器集成的性能明显优于四个标准预测模型中的三个。此外,我们对神经胶质瘤数据集进行全面的描述性分析,以识别相关的统计特征,并使用四种特征排序算法发现信息最丰富的属性。
    Accurate prediction and grading of gliomas play a crucial role in evaluating brain tumor progression, assessing overall prognosis, and treatment planning. In addition to neuroimaging techniques, identifying molecular biomarkers that can guide the diagnosis, prognosis and prediction of the response to therapy has aroused the interest of researchers in their use together with machine learning and deep learning models. Most of the research in this field has been model-centric, meaning it has been based on finding better performing algorithms. However, in practice, improving data quality can result in a better model. This study investigates a data-centric machine learning approach to determine their potential benefits in predicting glioma grades. We report six performance metrics to provide a complete picture of model performance. Experimental results indicate that standardization and oversizing the minority class increase the prediction performance of four popular machine learning models and two classifier ensembles applied on a low-imbalanced data set consisting of clinical factors and molecular biomarkers. The experiments also show that the two classifier ensembles significantly outperform three of the four standard prediction models. Furthermore, we conduct a comprehensive descriptive analysis of the glioma data set to identify relevant statistical characteristics and discover the most informative attributes using four feature ranking algorithms.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    背景:在临床研究的二元分类中,案例到类的不平衡分布以及二元因变量和独立变量子集之间的极端关联水平可能会产生重大的分类问题。这些关键问题,即阶级不平衡和完全分离,导致临床研究中分类不准确和结果有偏差。
    方法:为了处理类不平衡和完成分离问题,我们建议使用模糊逻辑回归框架进行二元分类。模糊逻辑回归结合了系数的三角模糊数的组合,输入,并输出并产生清晰的分类结果。由于模糊逻辑对不平衡和分离问题的更好处理,模糊逻辑回归框架显示出强大的分类性能。因此,提高了分类精度,降低临床研究患者的错误分类条件和偏颇见解的风险。
    结果:在具有临床数据集的十二个二元分类问题上评估了模糊逻辑回归模型的性能。该模型具有一贯的高灵敏度,特异性,F1,精度,和所有临床数据集的Mathew相关系数得分。没有证据表明数据集中存在的不平衡或分离会产生影响。此外,我们将模糊逻辑回归分类性能与经典逻辑回归的两个版本和文献中的六个不同的基准来源进行比较。这六个来源总共提供了十种不同的拟议方法,并且通过计算每种方法的相同分类性能分数集来进行比较。不平衡或分离会影响十分之七的方法。其余三个在各自的临床研究中产生更好的分类性能。然而,这些都优于模糊逻辑回归框架。
    结论:模糊逻辑回归显示了对不平衡和分离的强大表现,提供准确的预测,因此,在临床研究中对患者进行分类的信息见解。
    BACKGROUND: In binary classification for clinical studies, an imbalanced distribution of cases to classes and an extreme association level between the binary dependent variable and a subset of independent variables can create significant classification problems. These crucial issues, namely class imbalance and complete separation, lead to classification inaccuracy and biased results in clinical studies.
    METHODS: To deal with class imbalance and complete separation problems, we propose using a fuzzy logistic regression framework for binary classification. Fuzzy logistic regression incorporates combinations of triangular fuzzy numbers for the coefficients, inputs, and outputs and produces crisp classification results. The fuzzy logistic regression framework shows strong classification performance due to fuzzy logic\'s better handling of imbalance and separation issues. Hence, classification accuracy is improved, mitigating the risk of misclassified conditions and biased insights for clinical study patients.
    RESULTS: The performance of the fuzzy logistic regression model is assessed on twelve binary classification problems with clinical datasets. The model has consistently high sensitivity, specificity, F1, precision, and Mathew\'s correlation coefficient scores across all clinical datasets. There is no evidence of impact from the imbalance or separation that exists in the datasets. Furthermore, we compare the fuzzy logistic regression classification performance against two versions of classical logistic regression and six different benchmark sources in the literature. These six sources provide a total of ten different proposed methodologies, and the comparison occurs by calculating the same set of classification performance scores for each method. Either imbalance or separation impacts seven out of ten methodologies. The remaining three produce better classification performance in their respective clinical studies. However, these are all outperformed by the fuzzy logistic regression framework.
    CONCLUSIONS: Fuzzy logistic regression showcases strong performance against imbalance and separation, providing accurate predictions and, hence, informative insights for classifying patients in clinical studies.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    监测日常生活活动(ADL)在衡量和响应一个人管理其基本身体需求的能力方面起着重要作用。用于监视ADL的有效识别系统必须成功地识别自然活动,这些活动也以不频繁的间隔实际发生。然而,现有的系统主要侧重于识别更可分离的,受控活动类型或在活动发生更频繁的平衡数据集上进行训练。在我们的工作中,我们调查了将机器学习应用于从完全野外环境中收集的不平衡数据集的相关挑战.此分析表明,将提高召回率的预处理技术与提高精度的后处理技术相结合,可以为ADL监控等任务提供更理想的模型。在使用野外数据的独立于用户的评估中,这些技术产生了一个模型,该模型实现了基于事件的F1评分超过0.9的刷牙,梳理头发,走路,洗手。这项工作解决了机器学习中的基本挑战,这些挑战需要解决,以便这些系统能够被部署并在现实世界中可靠地工作。
    Monitoring activities of daily living (ADLs) plays an important role in measuring and responding to a person\'s ability to manage their basic physical needs. Effective recognition systems for monitoring ADLs must successfully recognize naturalistic activities that also realistically occur at infrequent intervals. However, existing systems primarily focus on either recognizing more separable, controlled activity types or are trained on balanced datasets where activities occur more frequently. In our work, we investigate the challenges associated with applying machine learning to an imbalanced dataset collected from a fully in-the-wild environment. This analysis shows that the combination of preprocessing techniques to increase recall and postprocessing techniques to increase precision can result in more desirable models for tasks such as ADL monitoring. In a user-independent evaluation using in-the-wild data, these techniques resulted in a model that achieved an event-based F1-score of over 0.9 for brushing teeth, combing hair, walking, and washing hands. This work tackles fundamental challenges in machine learning that will need to be addressed in order for these systems to be deployed and reliably work in the real world.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    本文讨论了腹腔镜手术图像的语义分割,特别强调用较少数量的观察结构的分割。作为这项研究的结果,提出了深度神经网络架构的调整参数,实现对手术场景中所有结构的鲁棒分割。具有五个编码器-解码器的U-Net体系结构(U-Net5ed),实现了SegNet-VGG19和采用不同主干的DeepLabv3+。进行了三个主要实验,使用整流线性单元(ReLU),高斯误差线性单位(GELU),和Swish激活功能。应用的损失函数包括交叉熵(CE),焦点损失(FL),特沃斯基损失(TL),骰子损失(DiL),交叉熵骰子损失(CEDL),和交叉熵特沃斯基损失(CETL)。比较了具有动量的随机梯度下降(SGDM)和自适应矩估计(Adam)优化器的性能。定性和定量证实,DeepLabv3+和U-Net5ed架构产生了最好的结果。具有ResNet-50主干的DeepLabv3+架构,Swish激活功能,和CETL损失函数报告平均准确度(MAcc)为0.976,平均交集(MIoU)为0.977。用较少数量的观察结果对结构进行语义分割,比如肝静脉,胆囊管,肝韧带,和血,验证了所获得的结果与所咨询的文献相比是非常有竞争力和有前途的。建议的选定参数在YOLOv9架构中进行了验证,与原始体系结构获得的结果相比,它显示了语义分割的改进。
    This article addresses the semantic segmentation of laparoscopic surgery images, placing special emphasis on the segmentation of structures with a smaller number of observations. As a result of this study, adjustment parameters are proposed for deep neural network architectures, enabling a robust segmentation of all structures in the surgical scene. The U-Net architecture with five encoder-decoders (U-Net5ed), SegNet-VGG19, and DeepLabv3+ employing different backbones are implemented. Three main experiments are conducted, working with Rectified Linear Unit (ReLU), Gaussian Error Linear Unit (GELU), and Swish activation functions. The applied loss functions include Cross Entropy (CE), Focal Loss (FL), Tversky Loss (TL), Dice Loss (DiL), Cross Entropy Dice Loss (CEDL), and Cross Entropy Tversky Loss (CETL). The performance of Stochastic Gradient Descent with momentum (SGDM) and Adaptive Moment Estimation (Adam) optimizers is compared. It is qualitatively and quantitatively confirmed that DeepLabv3+ and U-Net5ed architectures yield the best results. The DeepLabv3+ architecture with the ResNet-50 backbone, Swish activation function, and CETL loss function reports a Mean Accuracy (MAcc) of 0.976 and Mean Intersection over Union (MIoU) of 0.977. The semantic segmentation of structures with a smaller number of observations, such as the hepatic vein, cystic duct, Liver Ligament, and blood, verifies that the obtained results are very competitive and promising compared to the consulted literature. The proposed selected parameters were validated in the YOLOv9 architecture, which showed an improvement in semantic segmentation compared to the results obtained with the original architecture.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    皮肤病变分类在各种皮肤状况的早期发现和诊断中起着至关重要的作用。计算机辅助诊断技术的最新进展有助于及时干预,从而改善患者的预后,特别是在缺乏专业知识的农村社区。尽管在皮肤病检测中广泛采用了卷积神经网络(CNN),其有效性受到可公开获取的皮肤病变数据集的有限大小和数据不平衡的阻碍.在这种情况下,利用混合机器和深度学习(DL)技术,提出了一种两步分层二元分类方法。在国际皮肤成像合作(ISIC2017)数据集上进行的实验证明了分层方法在处理大类失衡方面的有效性。具体来说,采用DenseNet121(DNET)作为特征提取器和随机森林(RF)作为分类器产生了最有希望的结果,与纯深度学习模型(端到端DNET)相比,平衡的多类精度(BMA)达到91.07%,BMA为88.66%。RF集成在帮助DL解决有限数据学习的挑战方面比其他机器学习分类器表现出明显更高的效率。此外,实施的预测混合分层模型展示了增强的性能,同时显著减少了计算时间,表明其在实际应用中对皮肤病变分类的潜在效率。
    Skin lesion classification plays a crucial role in the early detection and diagnosis of various skin conditions. Recent advances in computer-aided diagnostic techniques have been instrumental in timely intervention, thereby improving patient outcomes, particularly in rural communities lacking specialized expertise. Despite the widespread adoption of convolutional neural networks (CNNs) in skin disease detection, their effectiveness has been hindered by the limited size and data imbalance of publicly accessible skin lesion datasets. In this context, a two-step hierarchical binary classification approach is proposed utilizing hybrid machine and deep learning (DL) techniques. Experiments conducted on the International Skin Imaging Collaboration (ISIC 2017) dataset demonstrate the effectiveness of the hierarchical approach in handling large class imbalances. Specifically, employing DenseNet121 (DNET) as a feature extractor and random forest (RF) as a classifier yielded the most promising results, achieving a balanced multiclass accuracy (BMA) of 91.07% compared to the pure deep-learning model (end-to-end DNET) with a BMA of 88.66%. The RF ensemble exhibited significantly greater efficiency than other machine-learning classifiers in aiding DL to address the challenge of learning with limited data. Furthermore, the implemented predictive hybrid hierarchical model demonstrated enhanced performance while significantly reducing computational time, indicating its potential efficiency in real-world applications for the classification of skin lesions.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

公众号