Convolutional neural network (CNN)

卷积神经网络 (CNN)
  • 文章类型: Journal Article
    在临床氟18氟脱氧葡萄糖PET/CT报告的基础上,评估领域适应对语言模型在预测五点多维尔分数方面的性能的影响。
    作者回顾性检索了2008年至2018年在威斯康星大学麦迪逊分校机构临床影像学数据库中进行的4542份氟脱氧葡萄糖PET/CT淋巴瘤检查的文本报告和图像。在这些总报告中,1664年,从报告中提取了多维尔分数,并用作培训标签。来自变压器(BERT)模型和初始化BERT模型的双向编码器表示BioClinicalBERT,拉德伯特,和RoBERTa通过使用蒙面语言建模进行预训练而适应了核医学领域。然后在五点Deauville分数预测的任务中,将这些域适应的模型与非域适应的版本进行比较。将语言模型与视觉模型进行了比较,多模态视觉语言模型,和核医学医生,七倍蒙特卡罗交叉验证。报告了用于准确性的均值和SDs,用配对t检验的P值。
    域自适应提高了所有语言模型的性能(P=.01)。例如,领域适应后,BERT从61.3%±2.9(SD)五级精度提高到65.7%±2.2(P=0.01)。领域改编的RoBERTa(名为DARoBERTa)表现最好,达到77.4%±3.4五级精度;该模型的性能与多模式对应物(称为多模式DARoBERTa)相似(77.2%±3.2),优于最佳仅视觉模型(48.1%±3.5,P≤.001)。对数据子集进行任务的医生具有66%的五级准确率。
    领域适应提高了大型语言模型在预测PET/CT报告中的Deauville分数方面的性能。关键词淋巴瘤,PET,PET/CT,迁移学习,无监督学习,卷积神经网络(CNN)核医学,多维尔,自然语言处理,多模态学习,人工智能,机器学习,语言建模补充材料可用于本文。©RSNA,202另见本期Abajian的评注。
    UNASSIGNED: To evaluate the impact of domain adaptation on the performance of language models in predicting five-point Deauville scores on the basis of clinical fluorine 18 fluorodeoxyglucose PET/CT reports.
    UNASSIGNED: The authors retrospectively retrieved 4542 text reports and images for fluorodeoxyglucose PET/CT lymphoma examinations from 2008 to 2018 in the University of Wisconsin-Madison institutional clinical imaging database. Of these total reports, 1664 had Deauville scores that were extracted from the reports and served as training labels. The bidirectional encoder representations from transformers (BERT) model and initialized BERT models BioClinicalBERT, RadBERT, and RoBERTa were adapted to the nuclear medicine domain by pretraining using masked language modeling. These domain-adapted models were then compared with the non-domain-adapted versions on the task of five-point Deauville score prediction. The language models were compared against vision models, multimodal vision-language models, and a nuclear medicine physician, with sevenfold Monte Carlo cross-validation. Means and SDs for accuracy are reported, with P values from paired t testing.
    UNASSIGNED: Domain adaptation improved the performance of all language models (P = .01). For example, BERT improved from 61.3% ± 2.9 (SD) five-class accuracy to 65.7% ± 2.2 (P = .01) following domain adaptation. Domain-adapted RoBERTa (named DA RoBERTa) performed best, achieving 77.4% ± 3.4 five-class accuracy; this model performed similarly to its multimodal counterpart (named Multimodal DA RoBERTa) (77.2% ± 3.2) and outperformed the best vision-only model (48.1% ± 3.5, P ≤ .001). A physician given the task on a subset of the data had a five-class accuracy of 66%.
    UNASSIGNED: Domain adaptation improved the performance of large language models in predicting Deauville scores in PET/CT reports.Keywords Lymphoma, PET, PET/CT, Transfer Learning, Unsupervised Learning, Convolutional Neural Network (CNN), Nuclear Medicine, Deauville, Natural Language Processing, Multimodal Learning, Artificial Intelligence, Machine Learning, Language Modeling Supplemental material is available for this article. © RSNA, 2023See also the commentary by Abajian in this issue.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    尽管衡量工人的生产率至关重要,测量每个工人的生产率是具有挑战性的,因为他们分散在不同的建筑工地。本文提出了一种基于惯性测量单元(IMU)和活动分类的生产率测量框架。利用两种深度学习算法和三种传感器组合来识别和分析该框架在砌体工作中的可行性。使用所提出的方法,使用具有多个传感器的卷积神经网络模型,可以以96.70%的最大准确度进行工人活动分类,使用带有单个传感器的长短期记忆(LSTM)模型,最低精度为72.11%。生产率的测量精度高达96.47%。这项研究的主要贡献是提出了一种对详细活动进行分类的方法,并探索了用于测量工人生产率的IMU传感器数量的影响。
    Although measuring worker productivity is crucial, the measurement of the productivity of each worker is challenging due to their dispersion across various construction jobsites. This paper presents a framework for measuring productivity based on an inertial measurement unit (IMU) and activity classification. Two deep learning algorithms and three sensor combinations were utilized to identify and analyze the feasibility of the framework in masonry work. Using the proposed method, worker activity classification could be performed with a maximum accuracy of 96.70% using the convolutional neural network model with multiple sensors, and a minimum accuracy of 72.11% using the long short-term memory (LSTM) model with a single sensor. Productivity could be measured with an accuracy of up to 96.47%. The main contributions of this study are the proposal of a method for classifying detailed activities and an exploration of the effect of the number of IMU sensors used in measuring worker productivity.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    深度学习对象检测模型已成功应用于开发计算机辅助诊断系统,以在结肠镜检查期间辅助息肉检测。这里,我们证明有必要包括阴性样本(I)减少息肉发现阶段的假阳性,通过包括具有可能混淆检测模型的伪像的图像(例如,医疗器械,水射流,粪便,血,相机过度靠近结肠壁,模糊的图像,等。)通常不包括在模型开发数据集中,和(Ii)正确估计模型的更现实的性能。通过重新训练我们先前开发的基于YOLOv3的检测模型,其数据集包括15%的其他非息肉图像,并带有各种伪影,我们能够在内部测试数据集中总体上提高其F1性能(从平均F1为0.869到0.893),现在包括这种类型的图像,以及包括非息肉图像的四个公共数据集(平均F1为0.695至0.722)。
    Deep learning object-detection models are being successfully applied to develop computer-aided diagnosis systems for aiding polyp detection during colonoscopies. Here, we evidence the need to include negative samples for both (i) reducing false positives during the polyp-finding phase, by including images with artifacts that may confuse the detection models (e.g., medical instruments, water jets, feces, blood, excessive proximity of the camera to the colon wall, blurred images, etc.) that are usually not included in model development datasets, and (ii) correctly estimating a more realistic performance of the models. By retraining our previously developed YOLOv3-based detection model with a dataset that includes 15% of additional not-polyp images with a variety of artifacts, we were able to generally improve its F1 performance in our internal test datasets (from an average F1 of 0.869 to 0.893), which now include such type of images, as well as in four public datasets that include not-polyp images (from an average F1 of 0.695 to 0.722).
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    导致许多人死亡的大流行之一是2019年冠状病毒病(COVID-19)。它首次出现在2019年底,许多死亡人数每天都在增加,直到现在。因此,COVID-19的早期诊断已成为一个突出的问题。此外,目前的诊断方法有几个缺点,需要进行新的调查以提高诊断性能。在本文中,执行一组阶段,比如收集数据,过滤和增强图像,提取特征,并对ECG图像进行分类。数据来自两个公开的ECG图像数据集,其中一份包含COVID心电图报告。一组预处理方法被应用于ECG图像,并且执行数据增强以基于类别来平衡ECG图像。基于卷积神经网络(CNN)的深度学习方法被执行用于特征提取。应用四种不同的预训练模型,例如Vgg16、Vgg19、ResNet-101和Xception。此外,Xception和临时卷积网络(TCN)的集合,它被命名为ECGConvnet,是提议的。最后,从以前的模型获得的结果被馈送到四个主要分类器。这些分类器是softmax,随机森林(RF),多层感知(MLP),和支持向量机(SVM)。前面的分类器用于评估所提出方法的诊断能力。分类方案基于五次交叉验证。提出了七个实验来评估ECGConvnet的性能。其中三个是多层次的,剩下的是二进制类诊断。七个实验中有六个诊断COVID-19患者。上述实验结果表明,与其他预训练模型相比,ECGConvnet具有最高的性能,与其他分类器相比,SVM分类器显示出更高的精度。基于SVM的ECGConvnet的结果精度为(99.74%,98.6%,在多类诊断任务上为99.1%)和(在二进制类诊断之一上为99.8%,而其余达到100%)。可以使用ECG数据开发基于深度学习的COVID自动诊断系统。
    One of the pandemics that have caused many deaths is the Coronavirus disease 2019 (COVID-19). It first appeared in late 2019, and many deaths are increasing day by day until now. Therefore, the early diagnosis of COVID-19 has become a salient issue. Additionally, the current diagnosis methods have several demerits, and a new investigation is required to enhance the diagnosis performance. In this paper, a set of phases are performed, such as collecting data, filtering and augmenting images, extracting features, and classifying ECG images. The data were obtained from two publicly available ECG image datasets, and one of them contained COVID ECG reports. A set of preprocessing methods are applied to the ECG images, and data augmentation is performed to balance the ECG images based on the classes. A deep learning approach based on a convolutional neural network (CNN) is performed for feature extraction. Four different pre-trained models are applied, such as Vgg16, Vgg19, ResNet-101, and Xception. Moreover, an ensemble of Xception and the temporary convolutional network (TCN), which is named ECGConvnet, is proposed. Finally, the results obtained from the former models are fed to four main classifiers. These classifiers are softmax, random forest (RF), multilayer perception (MLP), and support vector machine (SVM). The former classifiers are used to evaluate the diagnosis ability of the proposed methods. The classification scenario is based on fivefold cross-validation. Seven experiments are presented to evaluate the performance of the ECGConvnet. Three of them are multi-class, and the remaining are binary class diagnosing. Six out of seven experiments diagnose COVID-19 patients. The aforementioned experimental results indicated that ECGConvnet has the highest performance over other pre-trained models, and the SVM classifier showed higher accuracy in comparison with the other classifiers. The resulting accuracies from ECGConvnet based on SVM are (99.74%, 98.6%, 99.1% on the multi-class diagnosis tasks) and (99.8% on one of the binary-class diagnoses, while the remaining achieved 100%). It is possible to develop an automatic diagnosis system for COVID based on deep learning using ECG data.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    数字癌症双胞胎的发展依赖于在整个治疗过程中捕获单个癌症患者的高分辨率表示。我们的研究旨在通过将预测模型暴露于历史信息,从结构化放射学报告中提高对转移性疾病的检测。我们证明,自然语言处理(NLP)可以为计算机断层扫描(CT)报告的半监督分类生成更好的弱标签,当它通过患者的治疗历史连续报告时。来自纪念斯隆·凯特琳癌症中心的大约714,454份结构化放射学报告遵循标准化的部门结构化模板,用于模型开发,其中一部分报告用于验证。为了开发模型,对一部分报告进行了基本事实筛选:来自867例患者的肺转移数据集中的7,732份报告;来自315例患者的肝转移数据集中的2,777份报告;来自404例患者的肾上腺转移数据集中的4,107份报告.我们使用NLP从结构化文本报告中提取和编码重要特征,然后用于开发,火车,并验证模型。三种模型——一个简单的卷积神经网络(CNN),用注意力层增强的CNN,和循环神经网络(RNN)-被开发来分类转移性疾病的类型,并根据地面实况标签进行验证。模型使用来自患者的连续结构化文本放射学报告的特征来预测报告中转移性疾病的存在。单报告模型,以前开发用于分析一个报告而不是多个过去的报告,包括在内,并根据准确性比较所有四个模型的结果,精度,召回,和F1得分。最佳模型用于标记所有714,454个报告以生成转移图。我们的结果表明,与基于单一报告的预测相比,NLP模型可以从多个连续报告中提取癌症进展模式,并以更高的性能预测多个器官中转移性疾病的存在。它展示了一种有前途的自动化方法来标记大量的放射学报告,而无需以时间和成本有效的方式涉及人类专家,并能够随时间跟踪癌症进展。
    The development of digital cancer twins relies on the capture of high-resolution representations of individual cancer patients throughout the course of their treatment. Our research aims to improve the detection of metastatic disease over time from structured radiology reports by exposing prediction models to historical information. We demonstrate that Natural language processing (NLP) can generate better weak labels for semi-supervised classification of computed tomography (CT) reports when it is exposed to consecutive reports through a patient\'s treatment history. Around 714,454 structured radiology reports from Memorial Sloan Kettering Cancer Center adhering to a standardized departmental structured template were used for model development with a subset of the reports included for validation. To develop the models, a subset of the reports was curated for ground-truth: 7,732 total reports in the lung metastases dataset from 867 individual patients; 2,777 reports in the liver metastases dataset from 315 patients; and 4,107 reports in the adrenal metastases dataset from 404 patients. We use NLP to extract and encode important features from the structured text reports, which are then used to develop, train, and validate models. Three models-a simple convolutional neural network (CNN), a CNN augmented with an attention layer, and a recurrent neural network (RNN)-were developed to classify the type of metastatic disease and validated against the ground truth labels. The models use features from consecutive structured text radiology reports of a patient to predict the presence of metastatic disease in the reports. A single-report model, previously developed to analyze one report instead of multiple past reports, is included and the results from all four models are compared based on accuracy, precision, recall, and F1-score. The best model is used to label all 714,454 reports to generate metastases maps. Our results suggest that NLP models can extract cancer progression patterns from multiple consecutive reports and predict the presence of metastatic disease in multiple organs with higher performance when compared with a single-report-based prediction. It demonstrates a promising automated approach to label large numbers of radiology reports without involving human experts in a time- and cost-effective manner and enables tracking of cancer progression over time.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    Scene classification relying on images is essential in many systems and applications related to remote sensing. The scientific interest in scene classification from remotely collected images is increasing, and many datasets and algorithms are being developed. The introduction of convolutional neural networks (CNN) and other deep learning techniques contributed to vast improvements in the accuracy of image scene classification in such systems. To classify the scene from areal images, we used a two-stream deep architecture. We performed the first part of the classification, the feature extraction, using pre-trained CNN that extracts deep features of aerial images from different network layers: the average pooling layer or some of the previous convolutional layers. Next, we applied feature concatenation on extracted features from various neural networks, after dimensionality reduction was performed on enormous feature vectors. We experimented extensively with different CNN architectures, to get optimal results. Finally, we used the Support Vector Machine (SVM) for the classification of the concatenated features. The competitiveness of the examined technique was evaluated on two real-world datasets: UC Merced and WHU-RS. The obtained classification accuracies demonstrate that the considered method has competitive results compared to other cutting-edge techniques.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Sci-hub)

       PDF(Pubmed)

公众号