AUC, area under the receiver operating characteristic curve

AUC,接收器工作特性曲线下的面积
  • 文章类型: Editorial
    暂无摘要。
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    UNASSIGNED:多系统萎缩(MSA)患者轮椅依赖的预测因素尚不清楚。我们旨在探讨MSA患者早期轮椅依赖的预测因素,重点关注临床特征和血液生物标志物。
    未经评估:这是一项前瞻性队列研究。这项研究包括2014年1月至2019年12月期间诊断为MSA的患者。在2021年10月的截止日期,符合可能MSA诊断的患者被纳入分析。随机森林(RF)用于建立早期轮椅依赖的预测模型。准确性,灵敏度,特异性,和受试者工作特征曲线下面积(AUC)用于评估模型的性能。
    未经评估:总而言之,在RF模型中纳入了100例MSA患者,包括49例轮椅依赖患者和51例无轮椅依赖患者。轮椅依赖患者的基线血浆神经丝轻链(NFL)水平高于无轮椅依赖患者(P=0.037)。根据基尼指数,五个主要的预测因素是疾病持续时间,发病年龄,统一MSA评定量表(UMSARS)-II评分,NFL,和UMSARS-I得分,其次是C反应蛋白(CRP)水平,中性粒细胞与淋巴细胞比率(NLR),UMSARS-IV评分,症状发作,直立性低血压,性别,尿失禁,和诊断亚型。敏感性,特异性,准确度,RF模型的AUC为70.82%,74.55%,72.29%,和0.72。
    未经证实:除了临床特征,基线特征,包括NFL,CRP,NLR是MSA早期轮椅依赖的潜在预测生物标志物。这些发现为MSA早期干预试验提供了新的见解。
    UNASSIGNED: The predictive factors for wheelchair dependence in patients with multiple system atrophy (MSA) are unclear. We aimed to explore the predictive factors for early-wheelchair dependence in patients with MSA focusing on clinical features and blood biomarkers.
    UNASSIGNED: This is a prospective cohort study. This study included patients diagnosed with MSA between January 2014 and December 2019. At the deadline of October 2021, patients met the diagnosis of probable MSA were included in the analysis. Random forest (RF) was used to establish a predictive model for early-wheelchair dependence. Accuracy, sensitivity, specificity, and area under the receiver operating characteristic curve (AUC) were used to evaluate the performance of the model.
    UNASSIGNED: Altogether, 100 patients with MSA including 49 with wheelchair dependence and 51 without wheelchair dependence were enrolled in the RF model. Baseline plasma neurofilament light chain (NFL) levels were higher in patients with wheelchair dependence than in those without (P = 0.037). According to the Gini index, the five major predictive factors were disease duration, age of onset, Unified MSA Rating Scale (UMSARS)-II score, NFL, and UMSARS-I score, followed by C-reactive protein (CRP) levels, neutrophil-to-lymphocyte ratio (NLR), UMSARS-IV score, symptom onset, orthostatic hypotension, sex, urinary incontinence, and diagnosis subtype. The sensitivity, specificity, accuracy, and AUC of the RF model were 70.82 %, 74.55 %, 72.29 %, and 0.72, respectively.
    UNASSIGNED: Besides clinical features, baseline features including NFL, CRP, and NLR were potential predictive biomarkers of early-wheelchair dependence in MSA. These findings provide new insights into the trials regarding early intervention in MSA.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    未经证实:目前还没有确定的生物标志物用于抗VEGF治疗新生血管性年龄相关性黄斑变性(nAMD)的疗效和持久性。这项研究评估了基于放射学的定量OCT生物标志物,这些生物标志物可以预测抗VEGF治疗的反应和持久性。
    UNASSIGNED:使用机器学习(ML)分类器评估基线生物标志物以预测抗VEGF治疗的耐受性。
    未经评估:来自OSPREY研究的81名接受治疗的nAMD参与者,包括15名超级应答者(达到并维持视网膜液分辨率的患者)和66名非超级应答者(未达到或维持视网膜液分辨率的患者)。
    UNASSIGNED:从流体中提取了总共962个基于纹理的放射学特征,视网膜下高反射材料(SHRM),和OCT扫描的不同视网膜组织区室。前8个特点,通过最小冗余最大相关性特征选择方法选择,在交叉验证的方法中使用4个ML分类器进行评估,以区分2个患者组。还进行了基线和第3个月之间不同基于纹理的放射学描述符(δ-纹理特征)变化的纵向评估,以评估它们与治疗反应的关联。此外,8基线临床参数和基线OCT的组合,三角洲纹理特征,并通过交叉验证的方法评估了临床参数与治疗反应的相关性.
    UNASSIGNED:受试者工作特征曲线(AUC)下的交叉验证面积,准确度,灵敏度,并计算特异性以验证分类器的性能。
    UNASSIGNED:使用基于纹理的基线OCT特征,二次判别分析分类器的交叉验证AUC为0.75±0.09。基线和第3个月之间不同OCT区室内的δ-纹理特征产生0.78±0.08的AUC。基线临床参数视网膜下色素上皮体积和视网膜内液体积产生0.62±0.07的AUC。当所有的基线,delta,和临床特征相结合,分类器性能的统计显着提高(AUC,获得0.81±0.07)。
    UNASSIGNED:基于放射组学的OCT图像定量评估显示可区分nAMD中抗VEGF治疗的超应答者和非超应答者。发现基线流体和SHRM三角洲纹理特征在各组之间最具区别。
    UNASSIGNED: No established biomarkers currently exist for therapeutic efficacy and durability of anti-VEGF therapy in neovascular age-related macular degeneration (nAMD). This study evaluated radiomic-based quantitative OCT biomarkers that may be predictive of anti-VEGF treatment response and durability.
    UNASSIGNED: Assessment of baseline biomarkers using machine learning (ML) classifiers to predict tolerance to anti-VEGF therapy.
    UNASSIGNED: Eighty-one participants with treatment-naïve nAMD from the OSPREY study, including 15 super responders (patients who achieved and maintained retinal fluid resolution) and 66 non-super responders (patients who did not achieve or maintain retinal fluid resolution).
    UNASSIGNED: A total of 962 texture-based radiomic features were extracted from fluid, subretinal hyperreflective material (SHRM), and different retinal tissue compartments of OCT scans. The top 8 features, chosen by the minimum redundancy maximum relevance feature selection method, were evaluated using 4 ML classifiers in a cross-validated approach to distinguish between the 2 patient groups. Longitudinal assessment of changes in different texture-based radiomic descriptors (delta-texture features) between baseline and month 3 also was performed to evaluate their association with treatment response. Additionally, 8 baseline clinical parameters and a combination of baseline OCT, delta-texture features, and the clinical parameters were evaluated in a cross-validated approach in terms of association with therapeutic response.
    UNASSIGNED: The cross-validated area under the receiver operating characteristic curve (AUC), accuracy, sensitivity, and specificity were calculated to validate the classifier performance.
    UNASSIGNED: The cross-validated AUC by the quadratic discriminant analysis classifier was 0.75 ± 0.09 using texture-based baseline OCT features. The delta-texture features within different OCT compartments between baseline and month 3 yielded an AUC of 0.78 ± 0.08. The baseline clinical parameters sub-retinal pigment epithelium volume and intraretinal fluid volume yielded an AUC of 0.62 ± 0.07. When all the baseline, delta, and clinical features were combined, a statistically significant improvement in the classifier performance (AUC, 0.81 ± 0.07) was obtained.
    UNASSIGNED: Radiomic-based quantitative assessment of OCT images was shown to distinguish between super responders and non-super responders to anti-VEGF therapy in nAMD. The baseline fluid and SHRM delta-texture features were found to be most discriminating across groups.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    UNASSIGNED:开发一种客观分析常规白内障手术中可重复步骤的方法。
    未经评估:前瞻性研究;机器学习。
    UNASSIGNED:鉴定教师和培训生手术视频。
    UNASSIGNED:收集由教职员工或实习外科医生在眼科住院医师计划中连续6个月进行的白内障手术,并根据困难程度进行标记。现有的图像分类网络,ResNet152被微调用于白内障手术中的工具检测,以允许自动识别每个独特的手术器械。随后将各个显微镜视频帧窗口编码为矢量。检查了使用k倍用户输出交叉验证的向量编码与感知技能之间的关系。使用接受者工作特征曲线下面积(AUC)和分类准确度来评估算法。
    UASSIGNED:工具检测和技能评估的准确性。
    未经批准:总共,使用了391例连续白内障手术和209例常规病例。我们的模型实现了AUC范围从0.933到0.998的工具检测。对于技能分类,AUC为0.550(95%置信区间[CI],0.547-0.553),单个片段的准确率为54.3%(95%CI,53.9%-54.7%),一次手术的AUC为0.570(0.565-0.575),准确率为57.8%(56.8%-58.7%),AUC为0.692(0.659-0.758),单个用户在所有试验中的准确率为63.3%(56.8%-69.8%)。
    UNASSIGNED:我们的研究表明,机器学习可以准确,独立地识别视频中不同的白内障手术工具,这对于在一个步骤中比较工具的使用至关重要。然而,对于机器学习而言,准确区分整体和特定步骤技能以评估培训或专业知识水平更具挑战性。
    UNASSIGNED:作者对本文讨论的任何材料都没有专有或商业利益。
    UNASSIGNED: To develop a method for objective analysis of the reproducible steps in routine cataract surgery.
    UNASSIGNED: Prospective study; machine learning.
    UNASSIGNED: Deidentified faculty and trainee surgical videos.
    UNASSIGNED: Consecutive cataract surgeries performed by a faculty or trainee surgeon in an ophthalmology residency program over 6 months were collected and labeled according to degrees of difficulty. An existing image classification network, ResNet 152, was fine-tuned for tool detection in cataract surgery to allow for automatic identification of each unique surgical instrument. Individual microscope video frame windows were subsequently encoded as a vector. The relation between vector encodings and perceived skill using k-fold user-out cross-validation was examined. Algorithms were evaluated using area under the receiver operating characteristic curve (AUC) and the classification accuracy.
    UNASSIGNED: Accuracy of tool detection and skill assessment.
    UNASSIGNED: In total, 391 consecutive cataract procedures with 209 routine cases were used. Our model achieved an AUC ranging from 0.933 to 0.998 for tool detection. For skill classification, AUC was 0.550 (95% confidence interval [CI], 0.547-0.553) with an accuracy of 54.3% (95% CI, 53.9%-54.7%) for a single snippet, AUC was 0.570 (0.565-0.575) with an accuracy of 57.8% (56.8%-58.7%) for a single surgery, and AUC was 0.692 (0.659-0.758) with an accuracy of 63.3% (56.8%-69.8%) for a single user given all their trials.
    UNASSIGNED: Our research shows that machine learning can accurately and independently identify distinct cataract surgery tools in videos, which is crucial for comparing the use of the tool in a step. However, it is more challenging for machine learning to accurately differentiate overall and specific step skill to assess the level of training or expertise.
    UNASSIGNED: The author(s) have no proprietary or commercial interest in any materials discussed in this article.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    未经证实:早产儿视网膜病变(ROP)是与早产儿氧暴露有关的儿童失明的主要原因。由于氧气监测方案降低了需要治疗的ROP(TR-ROP)的发生率,目前尚不清楚氧暴露是否仍然是TR-ROP和侵袭性ROP(A-ROP)的相关危险因素,一个严重的,快速发展的ROP形式。这项概念验证研究的目的是使用电子健康记录(EHR)数据评估早期氧气暴露作为开发TR-ROP和A-ROP的预测变量。
    未经评估:回顾性队列研究。
    未经评估:在一个学术中心对240名婴儿进行ROP筛查。
    未经批准:对于每个婴儿,从EHR中手动提取氧饱和度和吸入氧分数(FiO2),直至月经后31周(PMA)。累积最小值,最大值,每周计算平均氧饱和度和FiO2。使用胎龄(GA)和30周PMA时的累积最小FiO2对随机森林模型进行5倍交叉验证,以鉴定发生TR-ROP的婴儿。对有或没有A-ROP的婴儿进行二级受试者工作特征(ROC)曲线分析,由于数量少,没有交叉验证。
    未经评估:对于每个型号,使用ROC曲线下面积(AUC)和精确召回曲线下面积(AUPRC)评分评估事件TR-ROP的交叉验证性能.对于A-ROP,我们在高敏操作点计算了AUC并评估了敏感性和特异性.
    未经批准:在包括的244名婴儿中,33开发的TR-ROP,其中5人开发了A-ROP。对于事故TR-ROP,在GA+累积最小FiO2(AUC=0.93±0.06;AUPRC=0.76±0.08)上训练的随机森林模型没有显著优于在单独GA上训练的模型(AUC=0.92±0.06[P=0.59];AUPRC=0.74±0.12[P=0.32]).仅使用氧的模型显示0.80±0.09的AUC。A-ROP的ROC分析发现AUC为0.92(95%置信区间,0.87-0.96)。
    未经评估:可以从EHR中提取氧气暴露,并将其量化为TR-ROP和A-ROP事件的危险因素。从EHR中提取可量化的临床特征可能有助于建立多种疾病的风险模型,并评估氧气暴露之间的复杂关系。拖放,和其他早产后遗症。
    UNASSIGNED: Retinopathy of prematurity (ROP) is a leading cause of childhood blindness related to oxygen exposure in premature infants. Since oxygen monitoring protocols have reduced the incidence of treatment-requiring ROP (TR-ROP), it remains unclear whether oxygen exposure remains a relevant risk factor for incident TR-ROP and aggressive ROP (A-ROP), a severe, rapidly progressing form of ROP. The purpose of this proof-of-concept study was to use electronic health record (EHR) data to evaluate early oxygen exposure as a predictive variable for developing TR-ROP and A-ROP.
    UNASSIGNED: Retrospective cohort study.
    UNASSIGNED: Two hundred forty-four infants screened for ROP at a single academic center.
    UNASSIGNED: For each infant, oxygen saturations and fraction of inspired oxygen (FiO2) were extracted manually from the EHR until 31 weeks postmenstrual age (PMA). Cumulative minimum, maximum, and mean oxygen saturation and FiO2 were calculated on a weekly basis. Random forest models were trained with 5-fold cross-validation using gestational age (GA) and cumulative minimum FiO2 at 30 weeks PMA to identify infants who developed TR-ROP. Secondary receiver operating characteristic (ROC) curve analysis of infants with or without A-ROP was performed without cross-validation because of small numbers.
    UNASSIGNED: For each model, cross-validation performance for incident TR-ROP was assessed using area under the ROC curve (AUC) and area under the precision-recall curve (AUPRC) scores. For A-ROP, we calculated AUC and evaluated sensitivity and specificity at a high-sensitivity operating point.
    UNASSIGNED: Of the 244 infants included, 33 developed TR-ROP, of which 5 developed A-ROP. For incident TR-ROP, random forest models trained on GA plus cumulative minimum FiO2 (AUC = 0.93 ± 0.06; AUPRC = 0.76 ± 0.08) were not significantly better than models trained on GA alone (AUC = 0.92 ± 0.06 [P = 0.59]; AUPRC = 0.74 ± 0.12 [P = 0.32]). Models using oxygen alone showed an AUC of 0.80 ± 0.09. ROC analysis for A-ROP found an AUC of 0.92 (95% confidence interval, 0.87-0.96).
    UNASSIGNED: Oxygen exposure can be extracted from the EHR and quantified as a risk factor for incident TR-ROP and A-ROP. Extracting quantifiable clinical features from the EHR may be useful for building risk models for multiple diseases and evaluating the complex relationships among oxygen exposure, ROP, and other sequelae of prematurity.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    UNASSIGNED:研究在基于OCT的血管造影成像期间调整眼放大倍数对结构-功能关系和青光眼检测的影响。
    未经评估:横断面研究。
    UNASSIGNED:共纳入96名健康对照参与者和90名开角型青光眼患者。
    UNASSIGNED:对对照组和患者组的每个患者的一只眼睛进行评估。包含黄斑血管密度(VD)和周围乳头VD的层来自扫描源OCT血管造影成像。使用Humphrey24-2测试测量标准自动视野法的平均灵敏度(MS)。用简单和部分相关系数评估了结构-功能关系。使用接受者工作特性曲线下面积(AUC)进行接受者工作特性分析以评估青光眼的诊断准确性。使用Bennett修改的Littmann公式调整眼睛放大倍数。
    UNASSIGNED:轴向长度与VD之间的关联,结构-功能关系,和青光眼检测有和没有放大校正。
    UNASSIGNED:在未进行放大校正的情况下,黄斑区的浅层与轴向长度没有显着相关(r=0.0011;P=0.99);但是,放大校正后与眼轴长度呈负相关(r=-0.22;P=0.028)。关于周围乳头区域的神经头层,观察到与没有放大校正的轴向长度呈负相关(r=-0.22;P=0.031);然而,这种显著的相关性随着放大校正而消失。黄斑浅层和周围乳头区域的神经头层与未经放大校正的Humphrey24-2MS值显着相关(分别为r=0.22和r=0.32);然而,这些相关性在放大倍数校正后没有改善(分别为r=0.20和r=0.33).浅层青光眼诊断准确性(AUC,0.63)和神经头层(AUC,0.70)无放大校正后没有改善(AUC,分别为0.62和0.69)。
    UNASSIGNED:调整眼睛放大倍数对于准确的VD测量很重要;但是,它可能不会显著影响结构-功能关系和青光眼的检测.
    UNASSIGNED: To investigate the effects of adjusting the ocular magnification during OCT-based angiography imaging on structure-function relationships and glaucoma detection.
    UNASSIGNED: Cross-sectional study.
    UNASSIGNED: A total of 96 healthy control participants and 90 patients with open-angle glaucoma were included.
    UNASSIGNED: One eye of each patient in the control group and the patient group was evaluated. The layers comprising the macula vascular density (VD) and circumpapillary VD were derived from swept-source OCT angiography imaging. The mean sensitivity (MS) of the standard automated perimetry was measured using the Humphrey 24-2 test. Structure-function relationships were evaluated with simple and partial correlation coefficients. A receiver operating characteristic analysis was performed to evaluate the diagnostic accuracy for glaucoma using the area under the receiver operating characteristic curve (AUC). Ocular magnification was adjusted using Littmann\'s formula modified by Bennett.
    UNASSIGNED: The association between the axial length and VD, structure-function relationships, and glaucoma detection with and without magnification correction.
    UNASSIGNED: The superficial layer of the macular region was not significantly correlated to the axial length without magnification correction (r = 0.0011; P = 0.99); however, it was negatively correlated to the axial length with magnification correction (r = -0.22; P = 0.028). Regarding the nerve head layer in the circumpapillary region, a negative correlation to the axial length without magnification correction was observed (r = -0.22; P = 0.031); however, this significant correlation disappeared with magnification correction. The superficial layer of the macula and the nerve head layer of the circumpapillary region were significantly correlated to Humphrey 24-2 MS values without magnification correction (r = 0.22 and r = 0.32, respectively); however, these correlations did not improve after magnification correction (r = 0.20 and r = 0.33, respectively). Glaucoma diagnostic accuracy in the superficial layer (AUC, 0.63) and nerve head layer (AUC, 0.70) without magnification correction did not improve after magnification correction (AUC, 0.62 and 0.69, respectively).
    UNASSIGNED: Adjustment of the ocular magnification is important for accurate VD measurements; however, it may not significantly impact structure-function relationships and glaucoma detection.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    UNASSIGNED:使用指示相对疾病严重程度的比较标签与来自早产儿视网膜病变(ROP)图像数据集的诊断类别标签,比较用于医学图像分类的训练神经网络的功效和效率。
    UNASSIGNED:诊断测试或技术的评估。
    UNASSIGNED:深度学习神经网络在接受诊断性ROP检查的患者获得的专家标记的广角视网膜图像上进行训练,作为ROP(i-ROP)队列研究的成像和信息学的一部分。
    UNASSIGNED:用来自2个数据集的ROP视网膜眼底图像中指示疾病严重程度的类别或比较标签训练神经网络。培训和验证后,在2个二元分类任务中的1个中,使用单独的测试数据集对所有网络进行评估:正常与异常或+与不+.
    UNASSIGNED:测量接收器工作特征曲线(AUC)值下的面积以评估网络性能。
    UNASSIGNED:给定相同数量的标签,相比之下,神经网络学习效率更高,在两个数据集的两个分类任务中生成显著更高的AUC。同样,给定相同数量的图像,在2个数据集中的1个中,比较学习在两个分类任务中开发的网络具有显著更高的AUC.在任一标签类型上训练的模型之间的效率和准确性差异随着训练集大小的增加而减小。
    UNASSIGNED:与类别标签相比,每个样本的比较标签信息更丰富。这些发现表明,在训练神经网络进行医学图像分类任务时,可以克服数据可变性和稀缺性的常见障碍。
    UNASSIGNED: To compare the efficacy and efficiency of training neural networks for medical image classification using comparison labels indicating relative disease severity versus diagnostic class labels from a retinopathy of prematurity (ROP) image dataset.
    UNASSIGNED: Evaluation of diagnostic test or technology.
    UNASSIGNED: Deep learning neural networks trained on expert-labeled wide-angle retinal images obtained from patients undergoing diagnostic ROP examinations obtained as part of the Imaging and Informatics in ROP (i-ROP) cohort study.
    UNASSIGNED: Neural networks were trained with either class or comparison labels indicating plus disease severity in ROP retinal fundus images from 2 datasets. After training and validation, all networks underwent evaluation using a separate test dataset in 1 of 2 binary classification tasks: normal versus abnormal or plus versus nonplus.
    UNASSIGNED: Area under the receiver operating characteristic curve (AUC) values were measured to assess network performance.
    UNASSIGNED: Given the same number of labels, neural networks learned more efficiently by comparison, generating significantly higher AUCs in both classification tasks across both datasets. Similarly, given the same number of images, comparison learning developed networks with significantly higher AUCs across both classification tasks in 1 of 2 datasets. The difference in efficiency and accuracy between models trained on either label type decreased as the size of the training set increased.
    UNASSIGNED: Comparison labels individually are more informative and more abundant per sample than class labels. These findings indicate a potential means of overcoming the common obstacle of data variability and scarcity when training neural networks for medical image classification tasks.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    UNASSIGNED:研究针对ETDRS网格量身定制的80秒多焦点瞳孔造影客观视野检查(mfPOP)测试的功效,以通过年龄相关性眼病研究(AREDS)严重程度等级诊断年龄相关性黄斑变性(AMD)。
    未经评估:诊断技术的评估。
    未经证实:我们比较了敏锐度的诊断能力,ETDRS网格视网膜厚度数据,新的80秒M18MFPOP测试,和两个宽视野6分钟的mfPOP测试(黄斑-P131,Widefield-P129)。M18刺激匹配分叉ETDRS网格区域的大小和形状,允许简单的结构-功能比较。M18、P129和P131刺激同时测试双眼。我们招募了34例早期AMD患者,平均±标准差(SD)年龄为72.6±7.06岁。M18和P129加上P131刺激有26和51名对照参与者,分别为73.1±8.17岁和72.1±5.83岁,分别。多焦点瞳孔成像客观视野检查使用了食品和药物管理局批准的客观现场分析仪(OFA;美国科南医疗公司)。
    UNASSIGNED:接受者操作员特征曲线(AUC)和Hedge\的g效应大小下的百分比面积。
    未经评估:Acuity和OCTETDRS网格厚度和体积为AREDS4级眼睛提供了合理的诊断能力(百分比AUC),分别为83.9±9.98%和90.2±6.32%(平均值±标准误差),分别,但不适用于疾病不太严重的眼睛。相比之下,M18刺激产生的AUC百分比从72.8±6.65%(区域2级)到92.9±3.93%(区域4级),全眼82.9±3.71%。对冲的g效应大小范围从0.84到2.32(大到巨大)。P131刺激的百分比AUC表现相似,而P129的百分比AUC表现较差。
    UNASSIGNED:快速客观的M18测试提供了与宽领域6分钟mfPOP测试相当的诊断能力。与敏锐度或OCTETDRS网格数据不同,OFA测试在AREDS1至3级眼睛中产生了合理的诊断能力。
    UNASSIGNED: To study the power of an 80-second multifocal pupillographic objective perimetry (mfPOP) test tailored to the ETDRS grid to diagnose age-related macular degeneration (AMD) by Age-Related Eye Disease Study (AREDS) severity grade.
    UNASSIGNED: Evaluation of a diagnostic technology.
    UNASSIGNED: We compared diagnostic power of acuity, ETDRS grid retinal thickness data, new 80-second M18 mfPOP test, and two wider-field 6-minute mfPOP tests (Macular-P131, Widefield-P129). The M18 stimuli match the size and shape of bifurcated ETDRS grid regions, allowing easy structure-function comparisons. M18, P129, and P131 stimuli test both eyes concurrently. We recruited 34 patients with early-stage AMD with a mean ± standard deviation (SD) age of 72.6 ± 7.06 years. The M18 and P129 plus P131 stimuli had 26 and 51 control participants, respectively with mean ± SD ages of 73.1 ± 8.17 years and 72.1 ± 5.83 years, respectively. Multifocal pupillographic objective perimetry testing used the Food and Drug Administration-cleared Objective FIELD Analyzer (OFA; Konan Medical USA).
    UNASSIGNED: Percentage area under the receiver operator characteristic curve (AUC) and Hedge\'s g effect size.
    UNASSIGNED: Acuity and OCT ETDRS grid thickness and volume produced reasonable diagnostic power (percentage AUC) for AREDS grade 4 eyes at 83.9 ± 9.98% and 90.2 ± 6.32% (mean ± standard error), respectively, but not for eyes with less severe disease. By contrast, M18 stimuli produced percentage AUCs from 72.8 ± 6.65% (AREDS grade 2) to 92.9 ± 3.93% (AREDS grade 4), and 82.9 ± 3.71% for all eyes. Hedge\'s g effect sizes ranged from 0.84 to 2.32 (large to huge). Percentage AUC for P131 stimuli performed similarly and for P129 performed somewhat less well.
    UNASSIGNED: The rapid and objective M18 test provided diagnostic power comparable with that of wider-field 6-minute mfPOP tests. Unlike acuity or OCT ETDRS grid data, OFA tests produced reasonable diagnostic power in AREDS grade 1 to 3 eyes.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    UNASSIGNED:开发基于图像的细菌和真菌角膜溃疡区分的计算机视觉模型,并将其性能与人类专家进行比较。
    UNASSIGNED:诊断性能的横截面比较。
    未经证实:急性,来自印度南部4个中心的培养证实的细菌或真菌性角膜炎。
    UNASSIGNED:使用手持摄像机的图像对五个卷积神经网络(CNN)进行了训练,这些图像是从印度南部经培养证实的角膜溃疡患者中收集的,作为2006年至2015年进行的临床试验的一部分。他们的表现在来自印度南部的2个搁置测试集(1个单中心和1个多中心)上进行了评估。十二名当地专家角膜专家对多中心测试集中的图像进行了远程解释,以实现与CNN性能的直接比较。
    UNASSIGNED:接收器工作特征曲线(AUC)下的面积单独和每组共同(即,CNN合奏和人类合奏)。
    未经评估:表现最好的CNN架构是MobileNet,在单中心测试集上达到0.86的AUC(其他CNN范围,0.68-0.84)和多中心测试集上的0.83(其他CNN范围,0.75-0.83)。多中心测试集上的专家人类AUC范围为0.42至0.79。CNN集合达到统计学上显著高于人类集合的AUC(0.84)(0.76;P<0.01)。CNNs对真菌(81%)和细菌(75%)溃疡的准确性相对较高。而人类对细菌性溃疡(88%)和真菌性溃疡(56%)的准确率相对较高。表现最好的CNN和表现最好的人的集合实现了0.87的最高AUC,尽管这在统计学上没有显著高于最好的CNN(0.83;P=0.17)或最好的人(0.79;P=0.09)。
    UNASSIGNED:与角膜专家相比,计算机视觉模型在识别角膜溃疡的潜在感染原因方面取得了超人的表现。表现最好的模型,MobileNet,在没有任何其他临床或历史信息的情况下,获得了0.83至0.86的AUC。这些研究结果表明,未来实施这些模型的潜力,使早期定向抗菌治疗能够控制感染性角膜炎。这可以改善视觉效果。正在进行其他研究,以将临床病史和专家意见纳入预测模型。
    UNASSIGNED: Develop computer vision models for image-based differentiation of bacterial and fungal corneal ulcers and compare their performance against human experts.
    UNASSIGNED: Cross-sectional comparison of diagnostic performance.
    UNASSIGNED: Patients with acute, culture-proven bacterial or fungal keratitis from 4 centers in South India.
    UNASSIGNED: Five convolutional neural networks (CNNs) were trained using images from handheld cameras collected from patients with culture-proven corneal ulcers in South India recruited as part of clinical trials conducted between 2006 and 2015. Their performance was evaluated on 2 hold-out test sets (1 single center and 1 multicenter) from South India. Twelve local expert cornea specialists performed remote interpretation of the images in the multicenter test set to enable direct comparison against CNN performance.
    UNASSIGNED: Area under the receiver operating characteristic curve (AUC) individually and for each group collectively (i.e., CNN ensemble and human ensemble).
    UNASSIGNED: The best-performing CNN architecture was MobileNet, which attained an AUC of 0.86 on the single-center test set (other CNNs range, 0.68-0.84) and 0.83 on the multicenter test set (other CNNs range, 0.75-0.83). Expert human AUCs on the multicenter test set ranged from 0.42 to 0.79. The CNN ensemble achieved a statistically significantly higher AUC (0.84) than the human ensemble (0.76; P < 0.01). CNNs showed relatively higher accuracy for fungal (81%) versus bacterial (75%) ulcers, whereas humans showed relatively higher accuracy for bacterial (88%) versus fungal (56%) ulcers. An ensemble of the best-performing CNN and best-performing human achieved the highest AUC of 0.87, although this was not statistically significantly higher than the best CNN (0.83; P = 0.17) or best human (0.79; P = 0.09).
    UNASSIGNED: Computer vision models achieved superhuman performance in identifying the underlying infectious cause of corneal ulcers compared with cornea specialists. The best-performing model, MobileNet, attained an AUC of 0.83 to 0.86 without any additional clinical or historical information. These findings suggest the potential for future implementation of these models to enable earlier directed antimicrobial therapy in the management of infectious keratitis, which may improve visual outcomes. Additional studies are ongoing to incorporate clinical history and expert opinion into predictive models.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    UNASSIGNED:开发和验证基于深度学习(DL)的自动化人工智能(AI)平台,用于使用基于镜片不透明度分类系统(LOCS)III的裂隙灯和逆透镜照片对白内障进行诊断和分级。
    UNASSIGNED:横断面研究,其中使用裂隙灯照片和逆透镜照片对卷积神经网络进行训练和测试。
    UNASISIGNED:来自596名患者的一千三百三十五个裂隙灯图像和637个逆向照明透镜图像。
    UNASSIGNED:由2名经过培训的分级人员使用LOCSIII对裂隙灯和后照镜头照片进行分级。图像数据集被标记并划分为训练,验证,和测试数据集。我们在AI领域使用4个关键策略对AI平台进行了训练和验证:(1)数据内部冗余信息的区域检测网络,(2)小数据集问题的数据扩充和迁移学习,(3)数据集偏差的广义交叉熵损失,(4)类不平衡问题的类平衡损失。AI平台的性能得到了3种AI算法的整合:ResNet18,WideResNet50-2和ResNext50。
    UNASSIGNED:基于诊断和LOCSIII的AI平台分级预测性能。
    UNASSIGNED:AI平台显示出强大的诊断性能(接收器工作特性曲线下的面积[AUC],0.9992[95%置信区间(CI),0.9986-0.9998]和0.9994[95%CI,0.9989-0.9998];准确性,98.82%[95%CI,97.7%-99.9%]和98.51%[95%CI,97.4%-99.6%])和基于LOCSIII的分级预测性能(AUC,0.9567[95%CI,0.9501-0.9633]和0.9650[95%CI,0.9509-0.9792];准确性,91.22%[95%CI,89.4%-93.0%]和90.26%[95%CI,88.6%-91.9%])使用裂隙灯照片进行核乳光(NO)和核颜色(NC),分别。对于皮质混浊(CO)和后囊膜下混浊(PSC),系统实现了高诊断性能(AUC,0.9680[95%CI,0.9579-0.9781]和0.9465[95%CI,0.9348-0.9582];准确性,96.21%[95%CI,94.4%-98.0%]和92.17%[95%CI,88.6%-95.8%])和良好的基于LOCSIII的分级预测性能(AUC,0.9044[95%CI,0.8958-0.9129]和0.9174[95%CI,0.9055-0.9295];准确性,91.33%[95%CI,89.7%-93.0%]和87.89%[95%CI,85.6%-90.2%])使用逆向照明图像。
    UNASSIGNED:我们基于DL的AI平台成功地在7级分类中对NO和NC以及6级分类中对CO和PSC进行了准确和精确的检测和分级,克服了医学数据库的局限性,例如训练数据很少或标签分布有偏差。
    UNASSIGNED: To develop and validate an automated deep learning (DL)-based artificial intelligence (AI) platform for diagnosing and grading cataracts using slit-lamp and retroillumination lens photographs based on the Lens Opacities Classification System (LOCS) III.
    UNASSIGNED: Cross-sectional study in which a convolutional neural network was trained and tested using photographs of slit-lamp and retroillumination lens photographs.
    UNASSIGNED: One thousand three hundred thirty-five slit-lamp images and 637 retroillumination lens images from 596 patients.
    UNASSIGNED: Slit-lamp and retroillumination lens photographs were graded by 2 trained graders using LOCS III. Image datasets were labeled and divided into training, validation, and test datasets. We trained and validated AI platforms with 4 key strategies in the AI domain: (1) region detection network for redundant information inside data, (2) data augmentation and transfer learning for the small dataset size problem, (3) generalized cross-entropy loss for dataset bias, and (4) class balanced loss for class imbalance problems. The performance of the AI platform was reinforced with an ensemble of 3 AI algorithms: ResNet18, WideResNet50-2, and ResNext50.
    UNASSIGNED: Diagnostic and LOCS III-based grading prediction performance of AI platforms.
    UNASSIGNED: The AI platform showed robust diagnostic performance (area under the receiver operating characteristic curve [AUC], 0.9992 [95% confidence interval (CI), 0.9986-0.9998] and 0.9994 [95% CI, 0.9989-0.9998]; accuracy, 98.82% [95% CI, 97.7%-99.9%] and 98.51% [95% CI, 97.4%-99.6%]) and LOCS III-based grading prediction performance (AUC, 0.9567 [95% CI, 0.9501-0.9633] and 0.9650 [95% CI, 0.9509-0.9792]; accuracy, 91.22% [95% CI, 89.4%-93.0%] and 90.26% [95% CI, 88.6%-91.9%]) for nuclear opalescence (NO) and nuclear color (NC) using slit-lamp photographs, respectively. For cortical opacity (CO) and posterior subcapsular opacity (PSC), the system achieved high diagnostic performance (AUC, 0.9680 [95% CI, 0.9579-0.9781] and 0.9465 [95% CI, 0.9348-0.9582]; accuracy, 96.21% [95% CI, 94.4%-98.0%] and 92.17% [95% CI, 88.6%-95.8%]) and good LOCS III-based grading prediction performance (AUC, 0.9044 [95% CI, 0.8958-0.9129] and 0.9174 [95% CI, 0.9055-0.9295]; accuracy, 91.33% [95% CI, 89.7%-93.0%] and 87.89% [95% CI, 85.6%-90.2%]) using retroillumination images.
    UNASSIGNED: Our DL-based AI platform successfully yielded accurate and precise detection and grading of NO and NC in 7-level classification and CO and PSC in 6-level classification, overcoming the limitations of medical databases such as few training data or biased label distribution.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

公众号