Convolutional neural network (CNN)

卷积神经网络 (CNN)
  • 文章类型: Journal Article
    对于空间物体探测任务,传统光学相机面临各种应用挑战,包括背光问题和昏暗的光线条件。作为一种新颖的光学相机,事件摄像机由于异步输出特性而具有高时间分辨率和高动态范围的优点,这为上述挑战提供了新的解决方案。然而,事件摄像机的异步输出特性使它们与为帧图像设计的常规目标检测方法不兼容。
    提出了用于处理事件摄像机数据的异步卷积存储器网络(ACMNet),以解决背光和昏暗空间物体检测的问题。ACMNet的关键思想是首先通过指数核函数用事件尖峰张量(EST)体素网格来表征异步事件流,然后使用前馈特征提取网络提取空间特征,并使用提出的卷积时空存储器模块ConvLSTM聚合时间特征,最后,实现了使用连续事件流的端到端对象检测。
    在Event_DVS_space7上进行了ACMNet和经典对象检测方法之间的比较实验,Event_DVS_space7是基于事件摄像机的大规模空间合成事件数据集。结果表明,ACMNet的性能优于其他ACMNet,mAP提高了12.7%,同时保持了处理速度。此外,事件摄像机在传统光学摄像机出现故障的背光和昏暗光线条件下仍然具有良好的性能。这项研究为在复杂的照明和运动条件下进行检测提供了一种新颖的可能性,强调事件相机在空间物体检测领域的优势。
    UNASSIGNED: For space object detection tasks, conventional optical cameras face various application challenges, including backlight issues and dim light conditions. As a novel optical camera, the event camera has the advantages of high temporal resolution and high dynamic range due to asynchronous output characteristics, which provides a new solution to the above challenges. However, the asynchronous output characteristic of event cameras makes them incompatible with conventional object detection methods designed for frame images.
    UNASSIGNED: Asynchronous convolutional memory network (ACMNet) for processing event camera data is proposed to solve the problem of backlight and dim space object detection. The key idea of ACMNet is to first characterize the asynchronous event streams with the Event Spike Tensor (EST) voxel grid through the exponential kernel function, then extract spatial features using a feed-forward feature extraction network, and aggregate temporal features using a proposed convolutional spatiotemporal memory module ConvLSTM, and finally, the end-to-end object detection using continuous event streams is realized.
    UNASSIGNED: Comparison experiments among ACMNet and classical object detection methods are carried out on Event_DVS_space7, which is a large-scale space synthetic event dataset based on event cameras. The results show that the performance of ACMNet is superior to the others, and the mAP is improved by 12.7% while maintaining the processing speed. Moreover, event cameras still have a good performance in backlight and dim light conditions where conventional optical cameras fail. This research offers a novel possibility for detection under intricate lighting and motion conditions, emphasizing the superior benefits of event cameras in the realm of space object detection.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    在靶向治疗前非侵入性检测肺腺癌患者的表皮生长因子受体(EGFR)突变状态仍然是一个挑战。这项研究旨在开发基于3维(3D)卷积神经网络(CNN)的深度学习模型,以使用计算机断层扫描(CT)图像预测EGFR突变状态。
    我们回顾性地从2个大型医疗中心收集了660名患者。根据医院来源将患者分为训练(n=528)和外部测试(n=132)组。CNN模型是以有监督的端到端方式训练的,并使用外部测试集评估其性能。为了比较CNN模型的性能,我们构建了1个临床和3个影像组学模型.此外,我们构建了一个综合模型,该模型结合了性能最高的影像组学和CNN模型.接收器工作特性(ROC)曲线用作每个模型的性能的主要量度。Delong测试用于比较不同模型之间的性能差异。
    与临床[训练集相比,曲线下面积(AUC)=69.6%,95%置信区间(CI),0.661-0.732;试验装置,AUC=68.4%,95%CI,0.609-0.752]和性能最高的影像组学模型(训练集,AUC=84.3%,95%CI,0.812-0.873;测试集,AUC=72.4%,95%CI,0.653-0.794)模型,CNN模型(训练集,AUC=94.3%,95%CI,0.920-0.961;测试集,AUC=94.7%,95%CI,0.894-0.978)对预测EGFR突变状态具有显著更好的预测性能。此外,与综合模型(训练集,AUC=95.7%,95%CI,0.942-0.971;测试集,AUC=87.4%,95%CI,0.820-0.924),CNN模型具有较好的稳定性。
    CNN模型在非侵入性预测肺腺癌患者的EGFR突变状态方面具有出色的性能,有望成为临床医生的辅助工具。
    UNASSIGNED: Noninvasively detecting epidermal growth factor receptor (EGFR) mutation status in lung adenocarcinoma patients before targeted therapy remains a challenge. This study aimed to develop a 3-dimensional (3D) convolutional neural network (CNN)-based deep learning model to predict EGFR mutation status using computed tomography (CT) images.
    UNASSIGNED: We retrospectively collected 660 patients from 2 large medical centers. The patients were divided into training (n=528) and external test (n=132) sets according to hospital source. The CNN model was trained in a supervised end-to-end manner, and its performance was evaluated using an external test set. To compare the performance of the CNN model, we constructed 1 clinical and 3 radiomics models. Furthermore, we constructed a comprehensive model combining the highest-performing radiomics and CNN models. The receiver operating characteristic (ROC) curves were used as primary measures of performance for each model. Delong test was used to compare performance differences between different models.
    UNASSIGNED: Compared with the clinical [training set, area under the curve (AUC) =69.6%, 95% confidence interval (CI), 0.661-0.732; test set, AUC =68.4%, 95% CI, 0.609-0.752] and the highest-performing radiomics models (training set, AUC =84.3%, 95% CI, 0.812-0.873; test set, AUC =72.4%, 95% CI, 0.653-0.794) models, the CNN model (training set, AUC =94.3%, 95% CI, 0.920-0.961; test set, AUC =94.7%, 95% CI, 0.894-0.978) had significantly better predictive performance for predicting EGFR mutation status. In addition, compared with the comprehensive model (training set, AUC =95.7%, 95% CI, 0.942-0.971; test set, AUC =87.4%, 95% CI, 0.820-0.924), the CNN model had better stability.
    UNASSIGNED: The CNN model has excellent performance in non-invasively predicting EGFR mutation status in patients with lung adenocarcinoma and is expected to become an auxiliary tool for clinicians.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    长期以来,维护通信网络的安全性一直是一个主要问题。由于物联网(IoT)等新通信架构的出现以及渗透技术的进步和复杂性,这个问题变得越来越重要。对于在基于物联网的网络中的使用,以前的入侵检测系统(IDS),通常使用集中式设计来识别威胁,现在是无效的。为了解决这些问题,这项研究提出了一种新的和协作的方法,物联网入侵检测,可能有助于解决某些当前的安全问题。建议的方法通过使用黑洞优化(BHO)来选择最能描述对象之间通信的最重要的属性。此外,提出了一种基于矩阵的网络通信特性描述方法。建议的入侵检测模型的输入由这两个特征集组成。所建议的技术使用软件定义网络(SDN)将网络分成多个子网。每个子网的监控由控制器节点完成,它使用卷积神经网络(PCNN)的并行组合来确定通过其子网的流量中是否存在安全威胁。所提出的方法还将多数投票方法用于控制器节点的协作,以便更准确地检测攻击。研究结果表明,与以前的方法相比,建议的合作策略可以检测NSLKDD和NSW-NB15数据集中的攻击,准确率为99.89%和97.72%,分别。这至少是0.6%的改善。
    Maintaining security in communication networks has long been a major concern. This issue has become increasingly crucial due to the emergence of new communication architectures like the Internet of Things (IoT) and the advancement and complexity of infiltration techniques. For usage in networks based on the Internet of Things, previous intrusion detection systems (IDSs), which often use a centralized design to identify threats, are now ineffective. For the resolution of these issues, this study presents a novel and cooperative approach to IoT intrusion detection that may be useful in resolving certain current security issues. The suggested approach chooses the most important attributes that best describe the communication between objects by using Black Hole Optimization (BHO). Additionally, a novel method for describing the network\'s matrix-based communication properties is put forward. The inputs of the suggested intrusion detection model consist of these two feature sets. The suggested technique splits the network into a number of subnets using the software-defined network (SDN). Monitoring of each subnet is done by a controller node, which uses a parallel combination of convolutional neural networks (PCNN) to determine the presence of security threats in the traffic passing through its subnet. The proposed method also uses the majority voting approach for the cooperation of controller nodes in order to more accurately detect attacks. The findings demonstrate that, in comparison to the prior approaches, the suggested cooperative strategy can detect assaults in the NSLKDD and NSW-NB15 datasets with an accuracy of 99.89 and 97.72 percent, respectively. This is a minimum 0.6 percent improvement.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    酸化会降低白云石的强度和质量,在某种程度上,损害隧道围岩的稳定性作为不利的地质边界。砂质白云石的沙化程度分级是复杂地理环境下隧道掘进等岩土工程项目面临的重要挑战之一。传统的定量测量物理参数或分析某些视觉特征的方法在实际使用中要么耗时要么不准确。为了解决这些问题,我们,第一次,将基于卷积神经网络(CNN)的图像分类方法引入到白云石沙化程度分类任务中。在这项研究中,我们通过建立包含5729张图像的大规模数据集做出了重大贡献,将砂质白云岩分为四种不同的沙化程度。这些图像是从中国CYWD项目玉溪段的隧道附近收集的。我们使用这个数据集进行了全面的分类实验。这些实验的结果表明了基于CNN的模型的开创性成就,实现了高达91.4%的令人印象深刻的准确率。这一成就强调了我们在创建该数据集方面的工作的先锋作用及其在复杂地理分析中的应用潜力。
    Sandification can degrade the strength and quality of dolomite, and to a certain extent, compromise the stability of a tunnel\'s surrounding rock as an unfavorable geological boundary. Sandification degree classification of sandy dolomite is one of the non-trivial challenges faced by geotechnical engineering projects such as tunneling in complex geographical environments. The traditional methods quantitatively measuring the physical parameters or analyzing some visual features are either time-consuming or inaccurate in practical use. To address these issues, we, for the first time, introduce the convolutional neural network (CNN)-based image classification methods into dolomite sandification degree classification task. In this study, we have made a significant contribution by establishing a large-scale dataset comprising 5729 images, classified into four distinct sandification degrees of sandy dolomite. These images were collected from the vicinity of a tunnel located in the Yuxi section of the CYWD Project in China. We conducted comprehensive classification experiments using this dataset. The results of these experiments demonstrate the groundbreaking achievement of CNN-based models, which achieved an impressive accuracy rate of up to 91.4%. This accomplishment underscores the pioneering role of our work in creating this dataset and its potential for applications in complex geographical analyses.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    在水下图像处理过程中,图像质量受光在水中的吸收和散射的影响,从而导致模糊和噪声等问题。因此,图像质量差是不可避免的。取得总体满意的研究成果,水下图像去噪至关重要。本文提出了一种水下图像去噪方法,名为HHDNet,旨在解决水下机器人摄影过程中环境干扰和技术限制引起的噪声问题。该方法利用双分支网络架构来处理高频和低频,结合了专门为去除水下图像中的高频突变噪声而设计的混合注意模块。使用高斯内核将输入图像分解为高频和低频分量。对于高频部分,具有混合注意力机制的全局上下文提取器(GCE)模块侧重于通过同时捕获局部细节和全局依赖性来去除高频突变信号。对于低频部分,考虑到较少的噪声信息,使用有效的残差卷积单元。实验结果表明,HHDNet能够有效地实现水下图像去噪,超越其他现有方法,不仅在去噪效果上,而且在保持计算效率上,因此,HHDNet在水下图像噪声去除方面提供了更大的灵活性。
    During underwater image processing, image quality is affected by the absorption and scattering of light in water, thus causing problems such as blurring and noise. As a result, poor image quality is unavoidable. To achieve overall satisfying research results, underwater image denoising is vital. This paper presents an underwater image denoising method, named HHDNet, designed to address noise issues arising from environmental interference and technical limitations during underwater robot photography. The method leverages a dual-branch network architecture to handle both high and low frequencies, incorporating a hybrid attention module specifically designed for the removal of high-frequency abrupt noise in underwater images. Input images are decomposed into high-frequency and low-frequency components using a Gaussian kernel. For the high-frequency part, a Global Context Extractor (GCE) module with a hybrid attention mechanism focuses on removing high-frequency abrupt signals by capturing local details and global dependencies simultaneously. For the low-frequency part, efficient residual convolutional units are used in consideration of less noise information. Experimental results demonstrate that HHDNet effectively achieves underwater image denoising tasks, surpassing other existing methods not only in denoising effectiveness but also in maintaining computational efficiency, and thus HHDNet provides more flexibility in underwater image noise removal.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    癫痫是全球最著名的神经系统疾病之一,导致个体突然癫痫发作并显著影响他们的生活质量。因此,迫切需要一种有效的方法来检测和预测癫痫发作,以减轻癫痫患者面临的风险。在本文中,提出了一种新的癫痫发作检测和预测方法,基于多类特征融合和卷积神经网络门控循环单元注意机制(CNN-GRU-AM)模型。最初,脑电图(EEG)信号通过离散小波变换(DWT)进行小波分解,产生六个子带。随后,从每个子带中提取时频域和非线性特征。最后,CNN-GRU-AM进一步提取特征并执行分类。CHB-MIT数据集用于验证所提出的方法。十倍交叉验证结果表明,我们的方法达到了99.24%和95.47%的灵敏度,特异性为99.51%和94.93%,准确率为99.35%和95.16%,在癫痫发作检测和预测任务中,AUC分别为99.34%和95.15%,分别。结果表明,本文提出的方法能够有效地实现癫痫发作的高精度检测和预测,以便提醒患者和医生及时采取防护措施。
    Epilepsy is one of the most well-known neurological disorders globally, leading to individuals experiencing sudden seizures and significantly impacting their quality of life. Hence, there is an urgent necessity for an efficient method to detect and predict seizures in order to mitigate the risks faced by epilepsy patients. In this paper, a new method for seizure detection and prediction is proposed, which is based on multi-class feature fusion and the convolutional neural network-gated recurrent unit-attention mechanism (CNN-GRU-AM) model. Initially, the Electroencephalography (EEG) signal undergoes wavelet decomposition through the Discrete Wavelet Transform (DWT), resulting in six subbands. Subsequently, time-frequency domain and nonlinear features are extracted from each subband. Finally, the CNN-GRU-AM further extracts features and performs classification. The CHB-MIT dataset is used to validate the proposed approach. The results of tenfold cross validation show that our method achieved a sensitivity of 99.24% and 95.47%, specificity of 99.51% and 94.93%, accuracy of 99.35% and 95.16%, and an AUC of 99.34% and 95.15% in seizure detection and prediction tasks, respectively. The results show that the method proposed in this paper can effectively achieve high-precision detection and prediction of seizures, so as to remind patients and doctors to take timely protective measures.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    植物病害显著影响作物产量和质量,对全球农业构成严重威胁。识别和分类这些疾病的过程通常是耗时的并且容易出错。本研究通过采用卷积神经网络和支持向量机(CNN-SVM)混合模型对四种经济上重要的农作物的疾病进行分类来解决这个问题:草莓,桃子,樱桃,和大豆。目标是对10类疾病进行分类,有六个患病班级和四个健康班级,对于这些作物,使用基于深度学习的CNN-SVM模型。几个预训练模型,包括VGG16、VGG19、DenseNet、盗梦空间,MobileNetV2,MobileNet,Xception,和ShuffleNet,也受过训练,实现精度范围从53.82%到98.8%。提出的模型,然而,平均准确率为99.09%。虽然所提出的模型的准确性与VGG16预训练模型相当,其显着较低的可训练参数数量使其更加高效和独特。这项研究证明了CNN-SVM模型在提高植物病害分类的准确性和效率方面的潜力。CNN-SVM模型由于其优越的性能指标而优于VGG16和其他模型。所提出的模型实现了99%的F1分数,曲线下面积(AUC)为99.98%,和99%的精度值,展示其功效。此外,使用梯度加权类激活图(Grad-CAM)技术生成类激活图,以提供检测到的疾病的视觉解释。创建了一个热图,以突出显示需要分类的区域,进一步验证模型的准确性和可解释性。
    Plant diseases significantly impact crop productivity and quality, posing a serious threat to global agriculture. The process of identifying and categorizing these diseases is often time-consuming and prone to errors. This research addresses this issue by employing a convolutional neural network and support vector machine (CNN-SVM) hybrid model to classify diseases in four economically important crops: strawberries, peaches, cherries, and soybeans. The objective is to categorize 10 classes of diseases, with six diseased classes and four healthy classes, for these crops using the deep learning-based CNN-SVM model. Several pre-trained models, including VGG16, VGG19, DenseNet, Inception, MobileNetV2, MobileNet, Xception, and ShuffleNet, were also trained, achieving accuracy ranges from 53.82% to 98.8%. The proposed model, however, achieved an average accuracy of 99.09%. While the proposed model\'s accuracy is comparable to that of the VGG16 pre-trained model, its significantly lower number of trainable parameters makes it more efficient and distinctive. This research demonstrates the potential of the CNN-SVM model in enhancing the accuracy and efficiency of plant disease classification. The CNN-SVM model was selected over VGG16 and other models due to its superior performance metrics. The proposed model achieved a 99% F1-score, a 99.98% Area Under the Curve (AUC), and a 99% precision value, demonstrating its efficacy. Additionally, class activation maps were generated using the Gradient Weighted Class Activation Mapping (Grad-CAM) technique to provide a visual explanation of the detected diseases. A heatmap was created to highlight the regions requiring classification, further validating the model\'s accuracy and interpretability.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    磁共振成像(MRI)软骨横向松弛时间(T2)反映了软骨组成,机械性能,和早期骨关节炎(OA)。T2分析需要软骨分割。在这项研究中,我们在临床上验证了前交叉韧带(ACL)损伤和健康膝盖在1.5特斯拉(T)的全自动T2分析。
    我们研究了71名参与者:20名ACL损伤患者,和22没有动态的膝盖不稳定,13进行手术重建,和16个健康对照。在基线和1年随访时获得矢状多回波自旋回波(MESE)MRI。手动分割股骨软骨;对来自同一扫描仪的MRI数据训练卷积神经网络(CNN)算法。
    71名参与者的自动分割与手动分割的骰子相似性系数(DSC)分别为0.83(股骨)和0.89(胫骨)。在自动分割(45.7±2.6ms)和手动分割(45.7±2.7ms)之间,股深T2相似(P=0.828),而表层T2通过自动分析略有高估(53.2±2.2vs.手动52.1±2.1ms;P<0.001)。深层的T2相关性为r=0.91-0.99,跨区域的表层的T2相关性为r=0.86-0.97。在股骨外侧的深层观察到1年内唯一具有统计学意义的T2增加[自动化与自动化的标准化反应平均值(SRM)=0.58手动分析为0.52;P<0.001]。ACL损伤组和健康参与者之间的基线/纵向T2值/变化没有相关差异,无论采用哪种分割方法。
    这项临床验证研究表明,从1.5T的MESE进行的自动化软骨T2分析在技术上是可行且准确的。可能需要更有效的3D序列和更长的观察间隔来检测ACL损伤诱导的关节不稳定性对软骨组成(T2)的影响。
    UNASSIGNED: Magnetic resonance imaging (MRI) cartilage transverse relaxation time (T2) reflects cartilage composition, mechanical properties, and early osteoarthritis (OA). T2 analysis requires cartilage segmentation. In this study, we clinically validate fully automated T2 analysis at 1.5 Tesla (T) in anterior cruciate ligament (ACL)-injured and healthy knees.
    UNASSIGNED: We studied 71 participants: 20 ACL-injured patients with, and 22 without dynamic knee instability, 13 with surgical reconstruction, and 16 healthy controls. Sagittal multi-echo-spin-echo (MESE) MRIs were acquired at baseline and 1-year follow-up. Femorotibial cartilage was segmented manually; a convolutional neural network (CNN) algorithm was trained on MRI data from the same scanner.
    UNASSIGNED: Dice similarity coefficients (DSCs) of automated versus manual segmentation in the 71 participants were 0.83 (femora) and 0.89 (tibiae). Deep femorotibial T2 was similar between automated (45.7±2.6 ms) and manual (45.7±2.7 ms) segmentation (P=0.828), whereas superficial layer T2 was slightly overestimated by automated analysis (53.2±2.2 vs. 52.1±2.1 ms for manual; P<0.001). T2 correlations were r=0.91-0.99 for deep and r=0.86-0.97 for superficial layers across regions. The only statistically significant T2 increase over 1 year was observed in the deep layer of the lateral femur [standardized response mean (SRM) =0.58 for automated vs. 0.52 for manual analysis; P<0.001]. There was no relevant difference in baseline/longitudinal T2 values/changes between the ACL-injured groups and healthy participants, with either segmentation method.
    UNASSIGNED: This clinical validation study suggests that automated cartilage T2 analysis from MESE at 1.5T is technically feasible and accurate. More efficient 3D sequences and longer observation intervals may be required to detect the impact of ACL injury induced joint instability on cartilage composition (T2).
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    心电图(ECG)已成为一种无处不在的诊断工具,用于识别和表征各种心血管疾病。可穿戴健康监测设备,配备了设备上的生物医学人工智能(AI)处理器,彻底改变了收购,分析,和心电图数据的解释。然而,这些系统需要具有灵活配置的人工智能处理器,便于携带,并在功耗和延迟方面展示实现各种功能的最佳性能。为了应对这些挑战,本研究提出了一种指令驱动的卷积神经网络(CNN)处理器。该处理器包含三个关键特征:(1)指令驱动的CNN处理器,以支持通用的基于ECG的应用。(2)同时考虑并行性和数据重用的处理元件(PE)阵列设计。(3)基于CORDIC算法的激活单元,支持Tanh和Sigmoid计算。本设计采用110nmCMOS工艺技术,占用1.35mm2的管芯面积,功耗为12.94µW。它已经被证明与两个典型的ECGAI应用,包括两类(即,正常/异常)分类和五类分类。提出的1-DCNN算法对两类分类具有97.95%的准确性,对五类分类具有97.9%的准确性,分别。
    Electrocardiography (ECG) has emerged as a ubiquitous diagnostic tool for the identification and characterization of diverse cardiovascular pathologies. Wearable health monitoring devices, equipped with on-device biomedical artificial intelligence (AI) processors, have revolutionized the acquisition, analysis, and interpretation of ECG data. However, these systems necessitate AI processors that exhibit flexible configuration, facilitate portability, and demonstrate optimal performance in terms of power consumption and latency for the realization of various functionalities. To address these challenges, this study proposes an instruction-driven convolutional neural network (CNN) processor. This processor incorporates three key features: (1) An instruction-driven CNN processor to support versatile ECG-based application. (2) A Processing element (PE) array design that simultaneously considers parallelism and data reuse. (3) An activation unit based on the CORDIC algorithm, supporting both Tanh and Sigmoid computations. The design has been implemented using 110 nm CMOS process technology, occupying a die area of 1.35 mm2 with 12.94 µW power consumption. It has been demonstrated with two typical ECG AI applications, including two-class (i.e., normal/abnormal) classification and five-class classification. The proposed 1-D CNN algorithm performs with a 97.95% accuracy for the two-class classification and 97.9% for the five-class classification, respectively.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    背景:胸部X线摄影是检测肋骨骨折的标准方法。我们的研究旨在开发一种人工智能(AI)模型,只有相对少量的训练数据,可以在胸片上识别肋骨骨折并准确标记其精确位置,从而实现与医疗专业人员相当的诊断准确性。方法:对于这项回顾性研究,我们使用540张标记为Detectron2的胸部X线照片(270张正常照片和270张肋骨骨折照片)开发了一个AI模型,该模型结合了一个更快的基于区域的卷积神经网络(R-CNN),增强了特征金字塔网络(FPN).评估了模型对X线照片进行分类和检测肋骨骨折的能力。此外,我们将模型的性能与12名医生的性能进行了比较,包括6名经委员会认证的麻醉师和6名住院医师,通过观察者性能测试。结果:关于AI模型的射线照相分类性能,灵敏度,特异性,受试者工作特征曲线下面积(AUROC)分别为0.87、0.83和0.89。在肋骨断裂检测性能方面,灵敏度,假阳性率,自由反应接收器工作特性(JAFROC)品质因数(FOM)分别为0.62、0.3和0.76。AI模型在观察者绩效测试中与12名医生中的11名和12名医生中的10名相比没有统计学上的显着差异,分别。结论:我们开发了一个在有限的数据集上训练的AI模型,该模型显示了与经验丰富的医生相当的肋骨骨折分类和检测性能。
    Background: Chest radiography is the standard method for detecting rib fractures. Our study aims to develop an artificial intelligence (AI) model that, with only a relatively small amount of training data, can identify rib fractures on chest radiographs and accurately mark their precise locations, thereby achieving a diagnostic accuracy comparable to that of medical professionals. Methods: For this retrospective study, we developed an AI model using 540 chest radiographs (270 normal and 270 with rib fractures) labeled for use with Detectron2 which incorporates a faster region-based convolutional neural network (R-CNN) enhanced with a feature pyramid network (FPN). The model\'s ability to classify radiographs and detect rib fractures was assessed. Furthermore, we compared the model\'s performance to that of 12 physicians, including six board-certified anesthesiologists and six residents, through an observer performance test. Results: Regarding the radiographic classification performance of the AI model, the sensitivity, specificity, and area under the receiver operating characteristic curve (AUROC) were 0.87, 0.83, and 0.89, respectively. In terms of rib fracture detection performance, the sensitivity, false-positive rate, and free-response receiver operating characteristic (JAFROC) figure of merit (FOM) were 0.62, 0.3, and 0.76, respectively. The AI model showed no statistically significant difference in the observer performance test compared to 11 of 12 and 10 of 12 physicians, respectively. Conclusions: We developed an AI model trained on a limited dataset that demonstrated a rib fracture classification and detection performance comparable to that of an experienced physician.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

公众号