VGG19

VGG19
  • 文章类型: Journal Article
    这项研究使用预处理的听觉脑干反应(ABR)图像数据评估了几种卷积神经网络(CNN)模型对患者听力损失进行分类的功效。具体来说,我们采用了六种CNN架构-VGG16,VGG19,DenseNet121,DenseNet-201,AlexNet,和InceptionV3-区分听力损失患者和听力正常患者。使用包含7990个预处理的ABR图像的数据集来评估这些模型的性能和准确性。对每个模型进行了系统测试,以确定其准确分类听力损失的能力。模型的比较分析侧重于准确性和计算效率的度量。结果表明,AlexNet模型表现出优异的性能,达到95.93%的精度。这项研究的结果表明,深度学习模型,特别是在这种情况下的AlexNet,具有使用ABR图数据自动诊断听力损失的巨大潜力。未来的工作将旨在完善这些模型,以提高其诊断准确性和效率。促进其在临床环境中的实际应用。
    This study evaluates the efficacy of several Convolutional Neural Network (CNN) models for the classification of hearing loss in patients using preprocessed auditory brainstem response (ABR) image data. Specifically, we employed six CNN architectures-VGG16, VGG19, DenseNet121, DenseNet-201, AlexNet, and InceptionV3-to differentiate between patients with hearing loss and those with normal hearing. A dataset comprising 7990 preprocessed ABR images was utilized to assess the performance and accuracy of these models. Each model was systematically tested to determine its capability to accurately classify hearing loss. A comparative analysis of the models focused on metrics of accuracy and computational efficiency. The results indicated that the AlexNet model exhibited superior performance, achieving an accuracy of 95.93%. The findings from this research suggest that deep learning models, particularly AlexNet in this instance, hold significant potential for automating the diagnosis of hearing loss using ABR graph data. Future work will aim to refine these models to enhance their diagnostic accuracy and efficiency, fostering their practical application in clinical settings.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    糖尿病足溃疡(DFU)对糖尿病患者构成重大威胁,经常导致截肢。早期发现DFU可以大大提高糖尿病患者的生存机会。这项工作介绍了FusionNet,一种新颖的多尺度特征融合网络,旨在使用多个预训练的卷积神经网络(CNN)算法准确区分DFU皮肤与健康皮肤。一个包含6963张来自不同患者的皮肤图像(3574张健康和3389张溃疡)的数据集被分为训练(6080张图像),验证(672张图像),和测试(211图像)集。最初,三种图像预处理技术——高斯滤波,中值滤波器,和运动模糊估计-用于消除不相关的,嘈杂,和模糊的数据。随后,利用三种预训练的CNN算法-DenseNet201,VGG19和NASNetMobile-从输入图像中提取高频特征。然后将这些特征输入到元调谐器模块中,以通过选择最具鉴别力的特征来预测DFU。统计检验,包括弗里德曼和方差分析(ANOVA),用于识别FusionNet和其他子网络之间的显著差异。最后,三种可解释的人工智能(XAI)算法-SHAP(Shapley加法扩张),LIME(本地可解释模型不可知解释),和Grad-CAM(梯度加权类激活映射)-被集成到FusionNet中,以提高透明度和可解释性。FusionNet分类器以99.05%的准确率获得了出色的分类结果,98.18%召回,100.00%精度,99.09%AUC,和99.08%的F1得分。我们相信,我们提出的FusionNet将成为医疗领域区分DFU与健康皮肤的有价值的工具。
    Diabetic foot ulcer (DFU) poses a significant threat to individuals affected by diabetes, often leading to limb amputation. Early detection of DFU can greatly improve the chances of survival for diabetic patients. This work introduces FusionNet, a novel multi-scale feature fusion network designed to accurately differentiate DFU skin from healthy skin using multiple pre-trained convolutional neural network (CNN) algorithms. A dataset comprising 6963 skin images (3574 healthy and 3389 ulcer) from various patients was divided into training (6080 images), validation (672 images), and testing (211 images) sets. Initially, three image preprocessing techniques - Gaussian filter, median filter, and motion blur estimation - were applied to eliminate irrelevant, noisy, and blurry data. Subsequently, three pre-trained CNN algorithms -DenseNet201, VGG19, and NASNetMobile - were utilized to extract high-frequency features from the input images. These features were then inputted into a meta-tuner module to predict DFU by selecting the most discriminative features. Statistical tests, including Friedman and analysis of variance (ANOVA), were employed to identify significant differences between FusionNet and other sub-networks. Finally, three eXplainable Artificial Intelligence (XAI) algorithms - SHAP (SHapley Additive exPlanations), LIME (Local Interpretable Model-agnostic Explanations), and Grad-CAM (Gradient-weighted Class Activation Mapping) - were integrated into FusionNet to enhance transparency and explainability. The FusionNet classifier achieved exceptional classification results with 99.05 % accuracy, 98.18 % recall, 100.00 % precision, 99.09 % AUC, and 99.08 % F1 score. We believe that our proposed FusionNet will be a valuable tool in the medical field to distinguish DFU from healthy skin.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    背景:肺癌是全球第二常见的癌症,每年有超过200万例新病例。早期识别将使医疗保健从业者更有效地处理它。计算机辅助检测系统的进步极大地影响了人类疾病的临床分析和决策。为此,机器学习和深度学习技术正在成功应用。由于几个优点,迁移学习已经成为基于图像数据的疾病检测的热点。
    方法:在这项工作中,我们通过堆叠三种不同的迁移学习模型来建立一种新颖的迁移学习模型(VER-Net),以使用肺部CT扫描图像检测肺癌。训练该模型以将CT扫描图像与四个肺癌类别映射。各种措施,如图像预处理,数据增强,和超参数调整,是为了提高VER-Net的功效。使用多分类胸部CT图像对所有模型进行训练和评估。
    结果:实验结果证实,与其他八种迁移学习模型相比,VER-Net的表现优于其他八种迁移学习模型。VER-Net得分91%,92%,91%,和91.3%时,测试的准确性,精度,召回,和F1得分,分别。与最先进的相比,VER-Net具有更好的准确性。
    结论:VER-Net不仅可有效用于肺癌检测,而且还可用于CT扫描图像可用的其他疾病。
    BACKGROUND: Lung cancer is the second most common cancer worldwide, with over two million new cases per year. Early identification would allow healthcare practitioners to handle it more effectively. The advancement of computer-aided detection systems significantly impacted clinical analysis and decision-making on human disease. Towards this, machine learning and deep learning techniques are successfully being applied. Due to several advantages, transfer learning has become popular for disease detection based on image data.
    METHODS: In this work, we build a novel transfer learning model (VER-Net) by stacking three different transfer learning models to detect lung cancer using lung CT scan images. The model is trained to map the CT scan images with four lung cancer classes. Various measures, such as image preprocessing, data augmentation, and hyperparameter tuning, are taken to improve the efficacy of VER-Net. All the models are trained and evaluated using multiclass classifications chest CT images.
    RESULTS: The experimental results confirm that VER-Net outperformed the other eight transfer learning models compared with. VER-Net scored 91%, 92%, 91%, and 91.3% when tested for accuracy, precision, recall, and F1-score, respectively. Compared to the state-of-the-art, VER-Net has better accuracy.
    CONCLUSIONS: VER-Net is not only effectively used for lung cancer detection but may also be useful for other diseases for which CT scan images are available.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    基于深度学习的对象检测方法已经在包括银行业在内的各个领域得到了应用,healthcare,电子治理,和学术界。近年来,从非结构化文档处理的不同场景或图像中进行文本检测和识别的研究工作受到了广泛的关注。本文的新颖之处在于详细讨论和实现了基于迁移学习的各种不同的印刷文本识别主干体系结构。在这篇研究文章中,作者比较了ResNet50、ResNet50V2、ResNet152V2、Inception、Xception,和VGG19骨干架构,具有预处理技术作为数据大小调整,归一化,以及标准OCRKaggle数据集上的噪声去除。Further,根据所达到的精度选择的前三个主干架构,然后执行超参数调谐以获得更准确的结果。Xception与ResNet相比表现良好,盗梦空间,VGG19,MobileNet架构通过实现高评估分数,准确性(98.90%)和最小损失(0.19)。根据该领域的现有研究,直到现在,在印刷或手写数据识别中使用的基于迁移学习的骨干体系结构在文献中没有得到很好的体现。我们将总的数据集分成80%用于训练,20%用于测试,然后分成不同的骨干架构模型,具有相同的纪元数,并发现Xception架构比其他架构实现了更高的准确度。此外,ResNet50V2模型的准确度(96.92%)高于ResNet152V2模型(96.34%).
    Object detection methods based on deep learning have been used in a variety of sectors including banking, healthcare, e-governance, and academia. In recent years, there has been a lot of attention paid to research endeavors made towards text detection and recognition from different scenesor images of unstructured document processing. The article\'s novelty lies in the detailed discussion and implementation of the various transfer learning-based different backbone architectures for printed text recognition. In this research article, the authors compared the ResNet50, ResNet50V2, ResNet152V2, Inception, Xception, and VGG19 backbone architectures with preprocessing techniques as data resizing, normalization, and noise removal on a standard OCR Kaggle dataset. Further, the top three backbone architectures selected based on the accuracy achieved and then hyper parameter tunning has been performed to achieve more accurate results. Xception performed well compared with the ResNet, Inception, VGG19, MobileNet architectures by achieving high evaluation scores with accuracy (98.90%) and min loss (0.19). As per existing research in this domain, until now, transfer learning-based backbone architectures that have been used on printed or handwritten data recognition are not well represented in literature. We split the total dataset into 80 percent for training and 20 percent for testing purpose and then into different backbone architecture models with the same number of epochs, and found that the Xception architecture achieved higher accuracy than the others. In addition, the ResNet50V2 model gave us higher accuracy (96.92%) than the ResNet152V2 model (96.34%).
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    目的:构建脑出血(sICH)后早期血肿扩大(HE)的临床非对比计算机断层扫描(NCCT)深度学习联合模型,并评估其预测性能。
    方法:纳入2017年1月至2022年12月西部战区总医院收治的254例原发性脑出血患者。根据血肿扩大超过33%或体积超过6ml的标准,将患者分为HE组和血肿未扩大(NHE)组。使用多个模型和10倍交叉验证方法来筛选最有价值的特征并建模预测HE的概率。曲线下面积(AUC)用于分析各模型对HE的预测效率。
    结果:将他们以8:2的比例随机分为204例训练集和50例测试集。临床影像学深层特征联合模型(22个特征)预测HE曲线下面积如下:临床NavieBayes模型AUC0.779,传统放射学逻辑回归(LR)模型AUC0.818,深度学习LR模型AUC0.873,临床NCCT深度学习多层感知器模型AUC0.921。
    结论:联合临床影像学深度学习模型对sICH患者早期HE有较高的预测作用,有助于临床个体化评估sICH患者早期HE的风险。
    OBJECTIVE: To construct a clinical noncontrastive computed tomography (NCCT) deep learning joint model for predicting early hematoma expansion (HE) after cerebral hemorrhage (sICH) and evaluate its predictive performance.
    METHODS: All 254 patients with primary cerebral hemorrhage from January 2017 to December 2022 in the General Hospital of the Western Theater Command were included. According to the criteria of hematoma enlargement exceeding 33% or the volume exceeding 6 ml, the patients were divided into the HE group and the hematoma non-enlargement (NHE) group. Multiple models and the 10-fold cross-validation method were used to screen the most valuable features and model the probability of predicting HE. The area under the curve (AUC) was used to analyze the prediction efficiency of each model for HE.
    RESULTS: They were randomly divided into a training set of 204 cases in an 8:2 ratio and 50 cases of the test set. The clinical imaging deep feature joint model (22 features) predicted the area under the curve of HE as follows: clinical Navie Bayes model AUC 0.779, traditional radiology logistic regression (LR) model AUC 0.818, deep learning LR model AUC 0.873, and clinical NCCT deep learning multilayer perceptron model AUC 0.921.
    CONCLUSIONS: The combined clinical imaging deep learning model has a high predictive effect for early HE in sICH patients, which is helpful for clinical individualized assessment of the risk of early HE in sICH patients.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    潜在低秩表示(LatLRR)已成为融合可见光和红外图像的一种重要方法。在这种方法中,图像被分解为三个基本成分:基础部分,显著部分,和稀疏的部分。目的是融合基础和显著特征以准确地重建图像。然而,现有的方法往往更侧重于将基础部分和显著部分结合起来,忽略稀疏成分的重要性,而我们主张将LatLRR图像分解生成的所有三个部分全面纳入图像融合过程,在这项研究中引入了一个新颖的命题。此外,卷积神经网络(CNN)技术与LatLRR的有效集成仍然具有挑战性,特别是在包含稀疏部分之后。这项研究利用了涉及加权平均的融合策略,求和,VGG19和ResNet50在各种组合中分析了稀疏部分引入后的融合性能。研究结果表明,通过在融合过程中包含稀疏零件,可以显着提高融合性能。建议的融合策略涉及采用深度学习技术来融合基础部分和稀疏部分,同时利用求和策略来融合显著部分。这些发现提高了基于LatLRR的方法的性能,并为增强提供了有价值的见解,导致图像融合领域的进步。
    Latent Low-Rank Representation (LatLRR) has emerged as a prominent approach for fusing visible and infrared images. In this approach, images are decomposed into three fundamental components: the base part, salient part, and sparse part. The aim is to blend the base and salient features to reconstruct images accurately. However, existing methods often focus more on combining the base and salient parts, neglecting the importance of the sparse component, whereas we advocate for the comprehensive inclusion of all three parts generated from LatLRR image decomposition into the image fusion process, a novel proposition introduced in this study. Moreover, the effective integration of Convolutional Neural Network (CNN) technology with LatLRR remains challenging, particularly after the inclusion of sparse parts. This study utilizes fusion strategies involving weighted average, summation, VGG19, and ResNet50 in various combinations to analyze the fusion performance following the introduction of sparse parts. The research findings show a significant enhancement in fusion performance achieved through the inclusion of sparse parts in the fusion process. The suggested fusion strategy involves employing deep learning techniques for fusing both base parts and sparse parts while utilizing a summation strategy for the fusion of salient parts. The findings improve the performance of LatLRR-based methods and offer valuable insights for enhancement, leading to advancements in the field of image fusion.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    深度学习和图像处理用于对乳腺肿瘤图像进行分类和分割,特别是在超声(美国)模式中,支持临床决策并提高医疗质量。然而,由于噪声和不同的成像方式,直接使用US图像可能具有挑战性。在这项研究中,我们开发了一个三步图像处理方案,涉及使用块匹配三维滤波技术进行斑点噪声滤波,感兴趣的区域突出显示,和RGB融合。该方法增强了深度学习模型的泛化,并获得了更好的性能。我们使用深度学习模型(VGG19)对三个数据集执行迁移学习:BUSI(780张图像),数据集B(162张图像),和KAIMRC(5693图像)。当使用五次交叉验证机制在BUSI和KAIMRC数据集上进行测试时,对于每个数据集,具有建议的预处理步骤的模型比没有预处理的模型表现更好。提出的图像处理方法提高了乳腺癌深度学习分类模型的性能。使用多个不同的数据集(私有和公共)来概括用于临床应用的模型。
    Deep learning and image processing are used to classify and segment breast tumor images, specifically in ultrasound (US) modalities, to support clinical decisions and improve healthcare quality. However, directly using US images can be challenging due to noise and diverse imaging modalities. In this study, we developed a three-step image processing scheme involving speckle noise filtering using a block-matching three-dimensional filtering technique, region of interest highlighting, and RGB fusion. This method enhances the generalization of deep-learning models and achieves better performance. We used a deep learning model (VGG19) to perform transfer learning on three datasets: BUSI (780 images), Dataset B (162 images), and KAIMRC (5693 images). When tested on the BUSI and KAIMRC datasets using a fivefold cross-validation mechanism, the model with the proposed preprocessing step performed better than without preprocessing for each dataset. The proposed image processing approach improves the performance of the breast cancer deep learning classification model. Multiple diverse datasets (private and public) were used to generalize the model for clinical application.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    口腔病变是口腔疾病的普遍表现,及时识别口腔病变对有效干预势在必行。幸运的是,深度学习算法已显示出自动化病变检测的巨大潜力。这项研究的主要目的是采用基于深度学习的图像分类算法来识别口腔病变。我们使用了三种深度学习模型,即VGG19,DeIT,和MobileNet,评估各种分类方法的有效性。为了评估模型的准确性和可靠性,我们使用了一个由口腔图片组成的数据集,其中包含两个不同的类别:良性和恶性病变。实验结果表明,VGG19和MobileNet几乎达到了100%的完美准确率,而DeIT的准确率略低,为98.73%。这项研究的结果表明,用于图片分类的深度学习算法在检测口腔病变方面表现出很高的有效性,VGG19和MobileNet达到100%,DeIT达到98.73%。具体来说,VGG19和MobileNet模型对这一特定任务表现出显著的适用性。
    Oral lesions are a prevalent manifestation of oral disease, and the timely identification of oral lesions is imperative for effective intervention. Fortunately, deep learning algorithms have shown great potential for automated lesion detection. The primary aim of this study was to employ deep learning-based image classification algorithms to identify oral lesions. We used three deep learning models, namely VGG19, DeIT, and MobileNet, to assess the efficacy of various categorization methods. To evaluate the accuracy and reliability of the models, we employed a dataset consisting of oral pictures encompassing two distinct categories: benign and malignant lesions. The experimental findings indicate that VGG19 and MobileNet attained an almost perfect accuracy rate of 100%, while DeIT achieved a slightly lower accuracy rate of 98.73%. The results of this study indicate that deep learning algorithms for picture classification demonstrate a high level of effectiveness in detecting oral lesions by achieving 100% for VGG19 and MobileNet and 98.73% for DeIT. Specifically, the VGG19 and MobileNet models exhibit notable suitability for this particular task.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    如今,由于人口的大幅扩张,自动疾病诊断已成为医学领域的重要角色。自动化的疾病诊断方法帮助临床医生诊断疾病,一致,和及时的结果,同时降低死亡率。视网膜脱离最近已成为最严重和最急性的眼部疾病之一,在全世界传播。因此,应实施自动化和最快的诊断模型,以便早期诊断视网膜脱离.本文介绍了一种新的最佳基平稳小波包变换和改进的VGG19-双向长短期记忆的混合方法,以使用视网膜眼底图像自动检测视网膜脱离。在本文中,利用最佳基平稳小波包变换进行图像分析,改进的VGG19-双向长短期记忆被用作深度特征提取器,然后通过自适应增强技术对获得的特征进行分类。实验结果表明,我们提出的方法获得了99.67%的灵敏度,95.95%特异性,98.21%精度,97.43%精度,98.54%F1得分,和0.9985AUC。该模型在目前可访问的数据库上获得了预期的结果,当其他RD图像变得可访问时,可以进一步增强。所提出的方法有助于眼科医生识别和轻松治疗RD患者。
    Nowadays, automated disease diagnosis has become a vital role in the medical field due to the significant population expansion. An automated disease diagnostic approach assists clinicians in the diagnosis of disease by giving exact, consistent, and prompt results, along with minimizing the mortality rate. Retinal detachment has recently emerged as one of the most severe and acute ocular illnesses, spreading worldwide. Therefore, an automated and quickest diagnostic model should be implemented to diagnose retinal detachment at an early stage. This paper introduces a new hybrid approach of best basis stationary wavelet packet transform and modified VGG19-Bidirectional long short-term memory to detect retinal detachment using retinal fundus images automatically. In this paper, the best basis stationary wavelet packet transform is utilized for image analysis, modified VGG19-Bidirectional long short-term memory is employed as the deep feature extractors, and then obtained features are classified through the Adaptive boosting technique. The experimental outcomes demonstrate that our proposed method obtained 99.67% sensitivity, 95.95% specificity, 98.21% accuracy, 97.43% precision, 98.54% F1-score, and 0.9985 AUC. The model obtained the intended results on the presently accessible database, which may be enhanced further when additional RD images become accessible. The proposed approach aids ophthalmologists in identifying and easily treating RD patients.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    脑中细胞的异常生长或脂肪团称为肿瘤。它们可以是健康的(正常的)或癌变,取决于细胞的结构。这可能会导致颅骨内压力增加,可能对大脑造成伤害甚至死亡.因此,诊断程序,如计算机断层扫描,磁共振成像,和正电子发射断层扫描,以及血液和尿液测试,用于识别脑肿瘤。然而,这些方法可能是劳动密集型的,有时会产生不准确的结果。而不是这些耗时的方法,采用深度学习模型是因为它们耗时较少,需要更便宜的设备,产生更准确的结果,并且很容易设置。在这项研究中,我们提出了一种基于迁移学习的方法,利用预训练的VGG-19模型。通过应用定制的卷积神经网络框架并将其与预处理方法相结合,这种方法得到了增强,包括规范化和数据扩充。对于培训和测试,我们提出的模型使用了数据集中80%和20%的图像,分别。我们提出的方法取得了显著的成功,准确率为99.43%,灵敏度为98.73%,特异性为97.21%。数据集,来自Kaggle,用于培训目的,由407张图片组成,包括257个描述脑肿瘤和150个没有肿瘤。这些模型可用于开发基于这些结果在CT图像中识别脑肿瘤的临床有用的解决方案。
    An abnormal growth or fatty mass of cells in the brain is called a tumor. They can be either healthy (normal) or become cancerous, depending on the structure of their cells. This can result in increased pressure within the cranium, potentially causing damage to the brain or even death. As a result, diagnostic procedures such as computed tomography, magnetic resonance imaging, and positron emission tomography, as well as blood and urine tests, are used to identify brain tumors. However, these methods can be labor-intensive and sometimes yield inaccurate results. Instead of these time-consuming methods, deep learning models are employed because they are less time-consuming, require less expensive equipment, produce more accurate results, and are easy to set up. In this study, we propose a method based on transfer learning, utilizing the pre-trained VGG-19 model. This approach has been enhanced by applying a customized convolutional neural network framework and combining it with pre-processing methods, including normalization and data augmentation. For training and testing, our proposed model used 80% and 20% of the images from the dataset, respectively. Our proposed method achieved remarkable success, with an accuracy rate of 99.43%, a sensitivity of 98.73%, and a specificity of 97.21%. The dataset, sourced from Kaggle for training purposes, consists of 407 images, including 257 depicting brain tumors and 150 without tumors. These models could be utilized to develop clinically useful solutions for identifying brain tumors in CT images based on these outcomes.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

公众号