prediction accuracy

预测精度
  • 文章类型: Journal Article
    这项研究旨在解决使用环境效益测绘和分析程序(BenMap)评估空气污染健康影响的准确性挑战。由于气象因子数据有限和污染物数据缺失造成的。通过采用数据增量策略和多种机器学习模型,这项研究探讨了数据量的影响,时间步长,以天津市几年来的数据为例,分析了气象因素对模型预测性能的影响。研究结果表明,增加训练数据量可以提高随机森林回归器(RF)和决策树回归器(DT)模型的性能。特别是预测CO,NO2和PM2.5。最佳预测时间步长因污染物而异,与DT模型实现最高的R2值(0.99)的CO和O3。综合多种气象因素,如大气压力,相对湿度,和露点温度,显著提高了模型精度。当使用三个气象因素时,该模型预测CO的R2为0.99,NO2、PM10、PM2.5和SO2。使用BenMap进行的健康影响评估表明,预测的全因死亡率和特定疾病死亡率与实际值高度一致,确认模型在评估空气污染对健康影响方面的准确性。例如,PM2.5的预测和实际全因死亡率均为3120;对于心血管疾病,两者都是1560年;对于呼吸系统疾病,都是780为了验证其通用性,该方法应用于成都,中国,利用几年的数据对PM2.5、CO、NO2、O3、PM10和SO2,结合大气压,相对湿度,和露点温度。该模型保持了优异的性能,确认其广泛的适用性。总的来说,我们得出的结论是,机器学习和基于BenMap的方法在预测空气污染物浓度和健康影响方面显示出很高的准确性和可靠性,为空气污染评估提供有价值的参考。
    This study aims to address accuracy challenges in assessing air pollution health impacts using Environmental Benefits Mapping and Analysis Program (BenMap), caused by limited meteorological factor data and missing pollutant data. By employing data increment strategies and multiple machine learning models, this research explores the effects of data volume, time steps, and meteorological factors on model prediction performance using several years of data from Tianjin City as an example. The findings indicate that increasing training data volume enhances the performance of Random Forest Regressor (RF) and Decision Tree Regressor (DT) models, especially for predicting CO, NO2, and PM2.5. The optimal prediction time step varies by pollutant, with the DT model achieving the highest R2 value (0.99) for CO and O3. Combining multiple meteorological factors, such as atmospheric pressure, relative humidity, and dew point temperature, significantly improves model accuracy. When using three meteorological factors, the model achieves an R2 of 0.99 for predicting CO, NO2, PM10, PM2.5, and SO2. Health impact assessments using BenMap demonstrated that the predicted all-cause mortality and specific disease mortalities were highly consistent with actual values, confirming the model\'s accuracy in assessing health impacts from air pollution. For instance, the predicted and actual all-cause mortality for PM2.5 were both 3120; for cardiovascular disease, both were 1560; and for respiratory disease, both were 780. To validate its generalizability, this method was applied to Chengdu, China, using several years of data for training and prediction of PM2.5, CO, NO2, O3, PM10, and SO2, incorporating atmospheric pressure, relative humidity, and dew point temperature. The model maintained excellent performance, confirming its broad applicability. Overall, we conclude that the machine learning and BenMap-based methods show high accuracy and reliability in predicting air pollutant concentrations and health impacts, providing a valuable reference for air pollution assessment.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    咖啡育种计划传统上依赖于多年来观察植物特性,一个缓慢而昂贵的过程。基因组选择(GS)提供了一种基于DNA的替代方法,可以更快地选择优质品种。堆叠集成学习(SEL)结合了多个模型,以实现更准确的选择。本研究探讨了SEL在咖啡育种中的潜力,旨在提高重要性状[产量(YL)的预测精度,水果总数(NF),叶子矿工侵扰(LM),阿拉比卡咖啡中的尾孢子虫病发病率(Cer)]。我们分析了来自195个个体的21,211个单核苷酸多态性(SNP)标记的基因分型数据。为了全面评估模型性能,我们采用了交叉验证(CV)方案。基因组最佳线性无偏预测(GBLUP),多元自适应回归样条(MARS),分位数随机森林(QRF),随机森林(RF)是基础学习者。对于SEL框架内的元学习器,探索了各种选择,包括岭回归,射频,GBLUP,和单一平均。SEL方法能够预测阿拉伯咖啡重要性状的预测能力(PA)。与所有基础学习方法获得的PA相比,SEL表现出更高的PA。PA相对于GBLUP的增益为87.44%(从最佳堆叠模型获得的PA与GBLUP之间的比率),37.83%,199.82%,YL为14.59%,NF,LM和Cer,分别。总的来说,SEL为GS提出了一种有前途的方法。通过组合来自多个模型的预测,SEL可以潜在地增强复杂性状的GS的PA。
    Coffee Breeding programs have traditionally relied on observing plant characteristics over years, a slow and costly process. Genomic selection (GS) offers a DNA-based alternative for faster selection of superior cultivars. Stacking Ensemble Learning (SEL) combines multiple models for potentially even more accurate selection. This study explores SEL potential in coffee breeding, aiming to improve prediction accuracy for important traits [yield (YL), total number of the fruits (NF), leaf miner infestation (LM), and cercosporiosis incidence (Cer)] in Coffea Arabica. We analyzed data from 195 individuals genotyped for 21,211 single-nucleotide polymorphism (SNP) markers. To comprehensively assess model performance, we employed a cross-validation (CV) scheme. Genomic Best Linear Unbiased Prediction (GBLUP), multivariate adaptive regression splines (MARS), Quantile Random Forest (QRF), and Random Forest (RF) served as base learners. For the meta-learner within the SEL framework, various options were explored, including Ridge Regression, RF, GBLUP, and Single Average. The SEL method was able to predict the predictive ability (PA) of important traits in Coffea Arabica. SEL presented higher PA compared with those obtained for all base learner methods. The gains in PA in relation to GBLUP were 87.44% (the ratio between the PA obtained from best Stacking model and the GBLUP), 37.83%, 199.82%, and 14.59% for YL, NF, LM and Cer, respectively. Overall, SEL presents a promising approach for GS. By combining predictions from multiple models, SEL can potentially enhance the PA of GS for complex traits.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    预测土壤重金属(SHM)含量对于了解城市居住区的SHM污染水平和指导减少污染的努力至关重要。然而,目前的研究表明,城市地区的SHM预测精度较低。因此,我们采用了深度学习方法(完全连接的深度神经网络)以及其他四种方法(多层感知器,径向基函数神经网络,多元逐步线性回归,和Kriging插值)来预测北京城市居住区的SHM含量,并展示了深度学习在提高预测精度方面的优势。我们发现评估的重金属含量(Cd,Cu,Pb,和Zn)与许多其他土壤理化性质和环境因子表现出显着相关性。铜的预测精度,Pb,不同方法的锌含量相对较高。值得注意的是,深度学习在预测四种重金属含量方面表现出相当大的优势,模型测试集的R2范围为0.75到0.91。与其他方法相比,根据不同的精度评估指标,深度学习实现了明显更高的预测精度(例如,深度学习显示,与其他方法相比,四种重金属的累积R2增加了53.16%至187.36%)。我们的研究表明,深度学习可以显着提高城市地区SHM内容预测的准确性,并且在具有复杂环境影响的城市居住区具有高度适用性。
    Predicting soil heavy metal (SHM) content is crucial for understanding SHM pollution levels in urban residential areas and guide efforts to reduce pollution. However, current research indicates low SHM prediction accuracy in urban areas. Therefore, we employed a deep learning method (fully connected deep neural network) alongside four other methods (muti-layer perceptron, radial basis function neural network, multiple stepwise linear regression, and Kriging interpolation) to predict SHM content in the urban residential areas of Beijing and demonstrated the strength of deep learning in improving prediction accuracy. We found the contents of the evaluated heavy metals (Cd, Cu, Pb, and Zn) exhibited significant correlations with numerous other soil physicochemical properties and environmental factors. The prediction accuracy for Cu, Pb, and Zn contents was relatively high across different methods. Notably, deep learning showed considerable strength in predicting the contents of the four heavy metals, with the R2 for the test set of the model ranging from 0.75 to 0.91. Compared to other methods, deep learning achieved markedly higher prediction accuracy according to different accuracy evaluation indicators (e.g., deep learning showed increases in the cumulative R2 of the four heavy metals ranging from 53.16 % to 187.36 % compared to other methods). Our study indicates that deep learning can significantly improve SHM content prediction accuracy in urban areas and is highly applicable in urban residential areas with complex environmental influences.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    本研究旨在比较使用基因组最佳线性无偏预测(GBLUP)方法估计的基因组估计育种值(GEBV)和结合来自全基因组关联研究(GWAS)的先前标记信息的GEBV估计的准确性。高原美利奴羊的断奶体重性状。旨在为提高基因组选择的准确性提供理论和技术支持。这项研究使用了1007只高原美利奴羊母羊,以3个月时的断奶体重为目标性状。将该人群随机分为两组。第一组用于GWAS分析,以识别重要的标记,最高的5%,前10%,前15%,选择前20%的标记作为先前标记信息。第二组用于估计遗传参数,并使用不同的先验标记信息比较GEBV预测的准确性。使用五倍交叉验证获得准确性。最后,两组均接受交叉验证.研究结果表明,断奶体重性状的遗传力,如使用GBLUP模型计算的,范围从0.122到0.394,相应的预测精度在0.075到0.228之间。通过合并来自GWAS的先前标记信息,遗传力提高到0.125至0.407。将来自GWAS结果的前5%至前20%显著SNP作为先验信息纳入GS显示出提高预测基因组育种价值的准确性的潜力。
    This study aims to compare the accuracy of genomic estimated breeding values (GEBV) estimated using a genomic best linear unbiased prediction (GBLUP) method and GEBV estimates incorporating prior marker information from a genome-wide association study (GWAS) for the weaning weight trait in highland Merino sheep. The objective is to provide theoretical and technical support for improving the accuracy of genomic selection. The study used a population of 1007 highland Merino ewes, with the weaning weight at 3 months as the target trait. The population was randomly divided into two groups. The first group was used for GWAS analysis to identify significant markers, and the top 5%, top 10%, top 15%, and top 20% markers were selected as prior marker information. The second group was used to estimate genetic parameters and compare the accuracy of GEBV predictions using different prior marker information. The accuracy was obtained using a five-fold cross-validation. Finally, both groups were subjected to cross-validation. The study\'s findings revealed that the heritability of the weaning weight trait, as calculated using the GBLUP model, ranged from 0.122 to 0.394, with corresponding prediction accuracies falling between 0.075 and 0.228. By incorporating prior marker information from GWAS, the heritability was enhanced to a range of 0.125 to 0.407. The inclusion of the top 5% to top 20% significant SNPs from GWAS results as prior information into GS showed potential for improving the accuracy of predicting genomic breeding value.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    使用来自广岛县的2494只日本黑牛进行了基因组预测,并通过气相色谱分析了单核苷酸多态性信息和单不饱和脂肪酸(MUFA)和油酸(C18:1)的表型数据。我们比较了四个模型的预测精度(A,加性遗传效应;AD,至于具有显性遗传效应的A;ADR,至于具有纯合性(ROH)效应的AD,由基于ROH的关系矩阵计算;和ADF,至于AD,用基于ROH的近交系数进行线性回归)。贝叶斯方法用于估计方差分量。MUFA和C18:1的狭义遗传力估计值分别为0.52-0.53和0.57;优势遗传变异的相应比例为0.04-0.07和0.04-0.05,ROH变异的比例为0.02。偏差信息标准值显示模型之间略有差异,模型提供了相似的预测精度。
    Genomic prediction was conducted using 2494 Japanese Black cattle from Hiroshima Prefecture and both single-nucleotide polymorphism information and phenotype data on monounsaturated fatty acid (MUFA) and oleic acid (C18:1) analyzed with gas chromatography. We compared the prediction accuracy for four models (A, additive genetic effects; AD, as for A with dominance genetic effects; ADR, as for AD with the runs of homozygosity (ROH) effects calculated by ROH-based relationship matrix; and ADF, as for AD with the ROH-based inbreeding coefficient of the linear regression). Bayesian methods were used to estimate variance components. The narrow-sense heritability estimates for MUFA and C18:1 were 0.52-0.53 and 0.57, respectively; the corresponding proportions of dominance genetic variance were 0.04-0.07 and 0.04-0.05, and the proportion of ROH variance was 0.02. The deviance information criterion values showed slight differences among the models, and the models provided similar prediction accuracy.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    基因组评估过程依赖于基因组水平的密集单核苷酸多态性(SNP)标记与数量性状基因座(QTL)之间的连锁不平衡假设。本研究的目的是评估四种频率方法,包括岭回归,最小绝对收缩和选择算子(LASSO),ElasticNet,基因组最佳线性无偏预测(GBLUP)和包括贝叶斯岭回归(BRR)在内的五种贝叶斯方法,贝叶斯A,贝叶斯LASSO,贝叶斯C,和贝叶斯B,在使用模拟数据的基因组选择中。基于统计显著性(p值)成对评估预测准确性之间的差异(即,t检验和Mann-WhitneyU检验)和实际意义(科恩的d效应大小)为此,数据是基于两种不同标记密度(整个基因组中的4000和8000)的情景进行模拟的。模拟数据包括一个有四个染色体的基因组,每个1摩根,其中100个随机分布的QTL和两个不同密度的均匀分布的SNP(1000和2000),在0.4的遗传力水平,被认为。对于除GBLUP外的频率论方法,正则化参数λ是使用五折交叉验证方法计算的。对于这两种情况,在频率论方法中,通过岭回归和GBLUP观察到最高的预测准确性。岭回归和GBLUP显示了最低和最高的偏差,分别。此外,在贝叶斯方法中,BayesB和BRR显示出最高和最低的预测精度,分别。贝叶斯LASSO记录了两种情况下的最低偏差,第一种和第二种情况下的最高偏差由BRR和贝叶斯B显示,分别。在这两种情况下的所有研究方法中,BayesB、LASSO和ElasticNet显示了最高和最低的精度,分别。不出所料,在GBLUP和BRR之间观察到最大的性能相似性(d=0.007,在第一种情况下,d=0.003,在第二种情况下)。从参数t和非参数Mann-WhitneyU检验获得的结果相似。在第一种和第二种情况下,在每个场景中所研究方法的性能之间进行36t检验,14(P<。001)和2(P<。05)比较显著,分别,这表明随着预测因子数量的增加,不同方法的性能差异减小。这是根据科恩的d效应大小证明的,因此,随着模型复杂性的增加,效应大小并没有被视为非常大。在将这些方法用于基因组评估之前,应通过交叉验证方法优化频率方法中的正则化参数。
    The genomic evaluation process relies on the assumption of linkage disequilibrium between dense single-nucleotide polymorphism (SNP) markers at the genome level and quantitative trait loci (QTL). The present study was conducted with the aim of evaluating four frequentist methods including Ridge Regression, Least Absolute Shrinkage and Selection Operator (LASSO), Elastic Net, and Genomic Best Linear Unbiased Prediction (GBLUP) and five Bayesian methods including Bayes Ridge Regression (BRR), Bayes A, Bayesian LASSO, Bayes C, and Bayes B, in genomic selection using simulation data. The difference between prediction accuracy was assessed in pairs based on statistical significance (p-value) (i.e., t test and Mann-Whitney U test) and practical significance (Cohen\'s d effect size) For this purpose, the data were simulated based on two scenarios in different marker densities (4000 and 8000, in the whole genome). The simulated data included a genome with four chromosomes, 1 Morgan each, on which 100 randomly distributed QTL and two different densities of evenly distributed SNPs (1000 and 2000), at the heritability level of 0.4, was considered. For the frequentist methods except for GBLUP, the regularization parameter λ was calculated using a five-fold cross-validation approach. For both scenarios, among the frequentist methods, the highest prediction accuracy was observed by Ridge Regression and GBLUP. The lowest and the highest bias were shown by Ridge Regression and GBLUP, respectively. Also, among the Bayesian methods, Bayes B and BRR showed the highest and lowest prediction accuracy, respectively. The lowest bias in both scenarios was registered by Bayesian LASSO and the highest bias in the first and the second scenario were shown by BRR and Bayes B, respectively. Across all the studied methods in both scenarios, the highest and the lowest accuracy were shown by Bayes B and LASSO and Elastic Net, respectively. As expected, the greatest similarity in performance was observed between GBLUP and BRR ( d = 0.007 , in the first scenario and d = 0.003 , in the second scenario). The results obtained from parametric t and non-parametric Mann-Whitney U tests were similar. In the first and second scenario, out of 36 t test between the performance of the studied methods in each scenario, 14 ( P < . 001 ) and 2 ( P < . 05 ) comparisons were significant, respectively, which indicates that with the increase in the number of predictors, the difference in the performance of different methods decreases. This was proven based on the Cohen\'s d effect size, so that with the increase in the complexity of the model, the effect size was not seen as very large. The regularization parameters in frequentist methods should be optimized by cross-validation approach before using these methods in genomic evaluation.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    适当的开花期是玉米育种的重要选择标准。它对玉米品种的生态适应性起着至关重要的作用。探讨开花时间的遗传基础,使用由379个多亲本DH系组成的关联组进行GWAS和GS分析。DH群体进行了几天的表型分析,以进行抽穗(DTT),花粉脱落天数(DTP),以及在不同环境中的天数(DTS)。遗传力为82.75%,86.09%,DTT为85.26%,DTP,和DTS,分别。使用FarmCPU模型的GWAS分析确定了分布在3、8、9和10号染色体上的10个单核苷酸多态性(SNP),这些多态性与开花时间相关的性状显着相关。BLINK模型的GWAS分析鉴定了分布在染色体1、3、8、9和10上的7个SNP,这些SNP与开花时间相关的性状显着相关。三个SNPs3_198946071、9_146646966和9_152140631显示多效效应,表明DTT之间存在显著的遗传相关性,DTP,和DTS。共检测到24个候选基因。从GWAS检测到100个显著相关的SNP,实现了相对较高的预测精度,最佳培训人口规模为70%。这项研究为更好地理解开花时间相关性状的遗传结构,并为GS提供了最佳策略。
    An appropriate flowering period is an important selection criterion in maize breeding. It plays a crucial role in the ecological adaptability of maize varieties. To explore the genetic basis of flowering time, GWAS and GS analyses were conducted using an associating panel consisting of 379 multi-parent DH lines. The DH population was phenotyped for days to tasseling (DTT), days to pollen-shedding (DTP), and days to silking (DTS) in different environments. The heritability was 82.75%, 86.09%, and 85.26% for DTT, DTP, and DTS, respectively. The GWAS analysis with the FarmCPU model identified 10 single-nucleotide polymorphisms (SNPs) distributed on chromosomes 3, 8, 9, and 10 that were significantly associated with flowering time-related traits. The GWAS analysis with the BLINK model identified seven SNPs distributed on chromosomes 1, 3, 8, 9, and 10 that were significantly associated with flowering time-related traits. Three SNPs 3_198946071, 9_146646966, and 9_152140631 showed a pleiotropic effect, indicating a significant genetic correlation between DTT, DTP, and DTS. A total of 24 candidate genes were detected. A relatively high prediction accuracy was achieved with 100 significantly associated SNPs detected from GWAS, and the optimal training population size was 70%. This study provides a better understanding of the genetic architecture of flowering time-related traits and provides an optimal strategy for GS.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    背景:森林遗传学家通常使用种源来解释其改良计划中的人口差异;然而,进口材料的历史记录可能不是非常精确或与先进分子技术衍生的遗传簇一致。这项研究的主要目的是评估基于标记的种群结构对与挪威云杉的生长和木材特性相关的遗传参数估计及其权衡的影响,要么将其作为固定效应(模型A)纳入,要么将其完全排除在分析之外(模型B)。
    结果:我们的结果表明,包含种群结构的模型显着降低了加性遗传变异的估计,导致狭义遗传力大幅降低。然而,这些模型大大提高了预测精度。这对生长和实木性能尤其重要,表明在所研究的性状中具有最高的群体遗传分化(QST)。此外,尽管模型之间的相关性模式仍然相似,对于将人口结构作为固定效应的模型,其幅度略低。这表明选择,在人群中一贯表现,与没有谱系限制的大量选择相比,可能受不利的遗传相关性影响较小。
    结论:我们得出的结论是,与忽略这种影响的模型相比,适当考虑人口结构的模型的结果更准确,偏差更小。这可能会对育种者和森林管理者产生实际影响,基于不精确选择的决策可能会给经济效率带来高风险。
    BACKGROUND: Forest geneticists typically use provenances to account for population differences in their improvement schemes; however, the historical records of the imported materials might not be very precise or well-aligned with the genetic clusters derived from advanced molecular techniques. The main objective of this study was to assess the impact of marker-based population structure on genetic parameter estimates related to growth and wood properties and their trade-offs in Norway spruce, by either incorporating it as a fixed effect (model-A) or excluding it entirely from the analysis (model-B).
    RESULTS: Our results indicate that models incorporating population structure significantly reduce estimates of additive genetic variance, resulting in substantial reduction of narrow-sense heritability. However, these models considerably improve prediction accuracies. This was particularly significant for growth and solid-wood properties, which showed to have the highest population genetic differentiation (QST) among the studied traits. Additionally, although the pattern of correlations remained similar across the models, their magnitude was slightly lower for models that included population structure as a fixed effect. This suggests that selection, consistently performed within populations, might be less affected by unfavourable genetic correlations compared to mass selection conducted without pedigree restrictions.
    CONCLUSIONS: We conclude that the results of models properly accounting for population structure are more accurate and less biased compared to those neglecting this effect. This might have practical implications for breeders and forest managers where, decisions based on imprecise selections can pose a high risk to economic efficiency.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    背景:识别高危患者并将其从初级保健医生(PCP)转诊给眼保健专业人员仍然是一个挑战。大约190万美国人由于未诊断或未治疗的眼科疾病而患有视力丧失。在眼科,人工智能(AI)用于预测青光眼进展,识别糖尿病视网膜病变(DR),并对眼部肿瘤进行分类;然而,AI尚未用于分类眼科转诊的初级保健患者。
    目的:本研究旨在构建和比较机器学习(ML)方法,适用于PCP的电子健康记录(EHR),能够将患者转诊给眼部护理专家。
    方法:访问Optum取消识别的EHR数据集,743,039例患者有5种主要视力状况(年龄相关性黄斑变性[AMD],视觉上显著的白内障,DR,青光眼,或眼表疾病[OSD])在年龄和性别上与无眼部疾病的743,039名对照完全匹配。每个患者的非眼科参数在142和182之间输入到5ML方法中:广义线性模型,L1正则化逻辑回归,随机森林,极端梯度提升(XGBoost),和J48决策树。比较每种病理的模型性能以选择最具预测性的算法。对每个结果的所有算法评估曲线下面积(AUC)。
    结果:XGBoost表现出最佳性能,显示,分别,对于视觉上有意义的白内障,预测准确性和AUC为78.6%(95%CI78.3%-78.9%)和0.878,77.4%(95%CI76.7%-78.1%)和0.858为渗出性AMD,非渗出性AMD为79.2%(95%CI78.8%-79.6%)和0.879,72.2%(95%CI69.9%-74.5%)和需要药物的OSD0.803,青光眼为70.8%(95%CI70.5%-71.1%)和0.785,85.0%(95%CI84.2%-85.8%),1型非增生性糖尿病视网膜病变(NPDR)为0.924,82.2%(95%CI80.4%-84.0%),1型增殖性糖尿病视网膜病变(PDR)为0.911,2型NPDR为81.3%(95%CI81.0%-81.6%)和0.891,2型PDR为82.1%(95%CI81.3%-82.9%)和0.900。
    结论:部署的5ML方法能够成功识别比值比(ORs)升高的患者,因此能够对患者进行分诊,对于眼病,从青光眼的2.4(95%CI2.4-2.5)到1型NPDR的5.7(95%CI5.0-6.4),平均OR为3.9。这些模型的应用可以使PCP更好地识别和分诊有可治疗眼科病理风险的患者。早期识别患有未识别的视力威胁疾病的患者可能会导致更早的治疗和减轻的经济负担。更重要的是,这样的分诊可以改善患者的生活。
    BACKGROUND: Identification and referral of at-risk patients from primary care practitioners (PCPs) to eye care professionals remain a challenge. Approximately 1.9 million Americans suffer from vision loss as a result of undiagnosed or untreated ophthalmic conditions. In ophthalmology, artificial intelligence (AI) is used to predict glaucoma progression, recognize diabetic retinopathy (DR), and classify ocular tumors; however, AI has not yet been used to triage primary care patients for ophthalmology referral.
    OBJECTIVE: This study aimed to build and compare machine learning (ML) methods, applicable to electronic health records (EHRs) of PCPs, capable of triaging patients for referral to eye care specialists.
    METHODS: Accessing the Optum deidentified EHR data set, 743,039 patients with 5 leading vision conditions (age-related macular degeneration [AMD], visually significant cataract, DR, glaucoma, or ocular surface disease [OSD]) were exact-matched on age and gender to 743,039 controls without eye conditions. Between 142 and 182 non-ophthalmic parameters per patient were input into 5 ML methods: generalized linear model, L1-regularized logistic regression, random forest, Extreme Gradient Boosting (XGBoost), and J48 decision tree. Model performance was compared for each pathology to select the most predictive algorithm. The area under the curve (AUC) was assessed for all algorithms for each outcome.
    RESULTS: XGBoost demonstrated the best performance, showing, respectively, a prediction accuracy and an AUC of 78.6% (95% CI 78.3%-78.9%) and 0.878 for visually significant cataract, 77.4% (95% CI 76.7%-78.1%) and 0.858 for exudative AMD, 79.2% (95% CI 78.8%-79.6%) and 0.879 for nonexudative AMD, 72.2% (95% CI 69.9%-74.5%) and 0.803 for OSD requiring medication, 70.8% (95% CI 70.5%-71.1%) and 0.785 for glaucoma, 85.0% (95% CI 84.2%-85.8%) and 0.924 for type 1 nonproliferative diabetic retinopathy (NPDR), 82.2% (95% CI 80.4%-84.0%) and 0.911 for type 1 proliferative diabetic retinopathy (PDR), 81.3% (95% CI 81.0%-81.6%) and 0.891 for type 2 NPDR, and 82.1% (95% CI 81.3%-82.9%) and 0.900 for type 2 PDR.
    CONCLUSIONS: The 5 ML methods deployed were able to successfully identify patients with elevated odds ratios (ORs), thus capable of patient triage, for ocular pathology ranging from 2.4 (95% CI 2.4-2.5) for glaucoma to 5.7 (95% CI 5.0-6.4) for type 1 NPDR, with an average OR of 3.9. The application of these models could enable PCPs to better identify and triage patients at risk for treatable ophthalmic pathology. Early identification of patients with unrecognized sight-threatening conditions may lead to earlier treatment and a reduced economic burden. More importantly, such triage may improve patients\' lives.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    确定“超额死亡率”可以比较国家之间以及随着时间的推移的灾害负担,从而评估缓解措施的成功。然而,关于2019年冠状病毒病(Covid-19)的辩论表明,根据方法和规格的不同,超额死亡率的计算差异很大。此外,通常不清楚“超额死亡率”的确切含义。我们将超额死亡率定义为超过预期的死亡人数,那是没有灾难性事件的。根据这个定义,我们使用了一种非常简约的计算方法,即对前几年的死亡数字进行线性外推,以确定新冠肺炎大流行期间的超额死亡率。但与大多数其他关于这个主题的文献不同,我们首先使用更大的历史数据集来评估和优化我们方法的规范,以识别和最小化估计误差和偏差。结果表明,文献中的超额死亡率经常被夸大。此外,如果这个价值在当时已经符合公众利益,他们在新冠肺炎之前的时期就会表现出相当大的超额死亡率。从这项研究及其发现中可以得出三个结论:(i)应首先根据过去的数字评估当前数字的所有计算方法。(ii)为避免警觉疲劳,应引入区分“通常波动”和“显著超额”的阈值。(三)统计局可以提供更现实的估计数。
    Determining \'excess mortality\' makes it possible to compare the burden of disasters between countries and over time, and thus also to evaluate the success of mitigation measures. However, the debate on coronavirus disease 2019 (Covid-19) has exposed that calculations of excess mortalities vary considerably depending on the method and its specification. Moreover, it is often unclear what exactly is meant by \'excess mortality\'. We define excess mortality as the excess over the number of deaths that would have been expected counter-factually, that is without the catastrophic event in question. Based on this definition, we use a very parsimonious calculation method, namely the linear extrapolation of death figures from previous years to determine the excess mortality during the Covid-19 pandemic. But unlike most other literature on this topic, we first evaluated and optimized the specification of our method using a larger historical data set in order to identify and minimize estimation errors and biases. The result shows that excess mortality rates in the literature are often inflated. Moreover, they would have exhibited considerable excess mortalities in the period before Covid-19, if this value had already been of public interest at that time. Three conclusions can be drawn from this study and its findings: (i) All calculation methods for current figures should first be evaluated against past figures. (ii) To avoid alarm fatigue, thresholds should be introduced which would differentiate between \'usual fluctuations\' and \'remarkable excess\'. (iii) Statistical offices could provide more realistic estimates.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

公众号