GoogleNet

GoogleNet
  • 文章类型: Journal Article
    分层是复合板中最重要和最危险的损坏之一。最近,许多论文提出了结构健康监测(SHM)技术的能力,用于研究具有各种形状和厚度深度的结构分层。然而,关于利用卷积神经网络(CNN)方法来自动化无损检测(NDT)技术数据库以识别分层大小和深度的研究很少。在本文中,提出了一种自动化系统,该系统能够区分原始结构和受损结构,并对具有不同深度的三类分层进行分类。该系统包括提出的CNN模型和兰姆波技术。在这项工作中,准备了一个单向复合板,该复合板在不同深度插入了三个分层样品,用于数值和实验研究。在数值部分,研究了导波的传播和与三个分层样品的相互作用,以观察分层深度如何影响分层区域的散射和捕获波。使用有效的超声导波技术对该数值研究进行了实验验证。该技术涉及压电晶片有源传感器(PWAS)和扫描激光多普勒振动计(SLDV)。数值和实验研究都表明,分层深度对捕获波的能量和分布有直接影响。从数值和实验研究中收集了三个不同的数据集,涉及数值波场图像数据集,实验波场图像数据集,和实验波数光谱图像数据集。这三个数据集与提出的CNN模型一起独立使用,以开发一个系统,该系统可以自动分类四个类别(原始类别和三个不同的分层类别)。所有三个数据集的结果显示了所提出的CNN模型以高精度预测分层深度的能力。使用GoogLeNetCNN验证了三个不同数据集的CNN模型结果。两种方法的结果都显示出极好的一致性。结果证明了波场图像和波数谱数据集用作CNN的输入数据以检测分层深度的能力。
    Delamination represents one of the most significant and dangerous damages in composite plates. Recently, many papers have presented the capability of structural health monitoring (SHM) techniques for the investigation of structural delamination with various shapes and thickness depths. However, few studies have been conducted regarding the utilization of convolutional neural network (CNN) methods for automating the non-destructive testing (NDT) techniques database to identify the delamination size and depth. In this paper, an automated system qualified for distinguishing between pristine and damaged structures and classifying three classes of delamination with various depths is presented. This system includes a proposed CNN model and the Lamb wave technique. In this work, a unidirectional composite plate with three samples of delamination inserted at different depths was prepared for numerical and experimental investigations. In the numerical part, the guided wave propagation and interaction with three samples of delamination were studied to observe how the delamination depth can affect the scattered and trapped waves over the delamination region. This numerical study was validated experimentally using an efficient ultrasonic guided waves technique. This technique involved piezoelectric wafer active sensors (PWASs) and a scanning laser Doppler vibrometer (SLDV). Both numerical and experimental studies demonstrate that the delamination depth has a direct effect on the trapped waves\' energy and distribution. Three different datasets were collected from the numerical and experimental studies, involving the numerical wavefield image dataset, experimental wavefield image dataset, and experimental wavenumber spectrum image dataset. These three datasets were used independently with the proposed CNN model to develop a system that can automatically classify four classes (pristine class and three different delamination classes). The results of all three datasets show the capability of the proposed CNN model for predicting the delamination depth with high accuracy. The proposed CNN model results of the three different datasets were validated using the GoogLeNet CNN. The results of both methods show an excellent agreement. The results proved the capability of the wavefield image and wavenumber spectrum datasets to be used as input data to the CNN for the detection of delamination depth.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    背景:检查DNA序列的功能和特征是一项极具挑战性的任务。说到人类基因组,由外显子和内含子组成,这项任务更具挑战性。人类外显子和内含子含有数百万到数十亿个核苷酸,这有助于在该序列中观察到的复杂性。考虑到基因组学的主题有多复杂,很明显,利用信号处理技术和深度学习工具建立一个强有力的预测模型,对人类基因组研究的发展是非常有帮助的。
    结果:使用频率混沌游戏表示用彩色图像表示人类外显子和内含子后,使用两个预训练的卷积神经网络模型(Resnet-50和GoogleNet)和一个建议的具有13个隐藏层的CNN模型对我们获得的图像进行分类。我们已经达到了92%的值的准确率为Resnet-50模型在大约7小时的执行时间,GoogleNet模型在2小时半的执行时间内的准确率值为91.5%。对于我们提出的CNN模型,在2h和37min内,我们的准确率达到91.6%。
    结论:我们提出的CNN模型在执行时间方面比Resnet-50模型更快。它能够稍微超过GoogleNet模型的准确率值。
    BACKGROUND: Examining functions and characteristics of DNA sequences is a highly challenging task. When it comes to the human genome, which is made up of exons and introns, this task is more challenging. Human exons and introns contain millions to billions of nucleotides, which contributes to the complexity observed in this sequences. Considering how complicated the subject of genomics is, it is obvious that using signal processing techniques and deep learning tools to build a strong predictive model can be very helpful for the development of the research of the human genome.
    RESULTS: After representing human exons and introns with color images using Frequency Chaos Game Representation, two pre-trained convolutional neural network models (Resnet-50 and GoogleNet) and a proposed CNN model having 13 hidden layers were used to classify our obtained images. We have reached a value of 92% for the accuracy rate for Resnet-50 model in about 7 h for the execution time, a value of 91.5% for the accuracy rate for the GoogleNet model in 2 h and a half for the execution time. For our proposed CNN model, we have reached 91.6% for the accuracy rate in 2 h and 37 min.
    CONCLUSIONS: Our proposed CNN model is faster than the Resnet-50 model in terms of execution time. It was able to slightly exceed the GoogleNet model for the accuracy rate value.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    多模式机器人音乐表演艺术领域由于其创新潜力而引起了极大的兴趣。传统的机器人在理解音乐表演中的情感和艺术表达方面面临局限性。因此,本文探讨了融合视觉和听觉感知的多模态机器人在音乐表演中的应用,以提高音乐表演的质量和艺术表现力。我们的方法涉及集成GRU(门控递归单元)和GoogLeNet模型进行情感分析。GRU模型处理音频数据并捕获音乐元素的时间动态,包括长期依赖,提取情感信息。GoogLeNet模型擅长图像处理,提取复杂的视觉细节和审美特征。这种协同作用加深了对音乐和视觉元素的理解,旨在产生更多情感共鸣和互动的机器人表演。实验结果证明了我们方法的有效性,显示出多模态机器人在音乐表现方面的显著改善。这些机器人,配备了我们的方法,提供高质量,有效唤起观众情感参与的艺术表演。在音乐表演中融合视听感知的多模态机器人丰富了艺术形式,并提供了多样化的人机交互。这项研究证明了多模式机器人在音乐表演中的潜力,促进技术与艺术的融合。它开辟了表演艺术和人机交互的新领域,提供独特和创新的体验。我们的发现为表演艺术领域多模式机器人的发展提供了宝贵的见解。
    The field of multimodal robotic musical performing arts has garnered significant interest due to its innovative potential. Conventional robots face limitations in understanding emotions and artistic expression in musical performances. Therefore, this paper explores the application of multimodal robots that integrate visual and auditory perception to enhance the quality and artistic expression in music performance. Our approach involves integrating GRU (Gated Recurrent Unit) and GoogLeNet models for sentiment analysis. The GRU model processes audio data and captures the temporal dynamics of musical elements, including long-term dependencies, to extract emotional information. The GoogLeNet model excels in image processing, extracting complex visual details and aesthetic features. This synergy deepens the understanding of musical and visual elements, aiming to produce more emotionally resonant and interactive robot performances. Experimental results demonstrate the effectiveness of our approach, showing significant improvements in music performance by multimodal robots. These robots, equipped with our method, deliver high-quality, artistic performances that effectively evoke emotional engagement from the audience. Multimodal robots that merge audio-visual perception in music performance enrich the art form and offer diverse human-machine interactions. This research demonstrates the potential of multimodal robots in music performance, promoting the integration of technology and art. It opens new realms in performing arts and human-robot interactions, offering a unique and innovative experience. Our findings provide valuable insights for the development of multimodal robots in the performing arts sector.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    开发有效的方法来推断由众多表情组成的不同面孔之间的关系或在不同时间在同一面孔上的关系(例如,疾病进展)是影像学相关研究中的一个悬而未决的问题。在这项研究中,我们提出了一种新颖的面部特征提取方法,表征,基于经典计算机视觉和深度学习的识别,更具体地说,卷积神经网络。
    我们描述了名为FRetrAival(FRAI)的混合面部表征系统,它是GoogleNet和AlexNet神经网络(NN)模型的混合。通过FRAI网络分析的图像通过计算机视觉技术进行预处理,例如基于梯度的定向算法,该算法只能从任何类型的图片中提取人脸区域。使用对齐人脸数据集(AFD)来训练和测试用于提取图像特征的FRAI解决方案。野生(LFW)保持数据集中的标记面已用于外部验证。
    总的来说,与以前的技术相比,我们的方法通过产生最大精度,在k-最近邻(KNN)上显示出更好的结果,召回,F1和F2得分值(92.00、92.66、92.33和92.52%,分别)对于AFD和(每个变量为95.00%)对于LFW数据集,它们被用作训练和测试数据集。FRAI模型可能会用于医疗保健和犯罪学以及许多其他应用中,在这些应用中,快速识别特定识别目标的指纹等面部特征非常重要。
    UNASSIGNED: Developing efficient methods to infer relations among different faces consisting of numerous expressions or on the same face at different times (e.g., disease progression) is an open issue in imaging related research. In this study, we present a novel method for facial feature extraction, characterization, and identification based on classical computer vision coupled with deep learning and, more specifically, convolutional neural networks.
    UNASSIGNED: We describe the hybrid face characterization system named FRetrAIval (FRAI), which is a hybrid of the GoogleNet and the AlexNet Neural Network (NN) models. Images analyzed by the FRAI network are preprocessed by computer vision techniques such as the oriented gradient-based algorithm that can extract only the face region from any kind of picture. The Aligned Face dataset (AFD) was used to train and test the FRAI solution for extracting image features. The Labeled Faces in the Wild (LFW) holdout dataset has been used for external validation.
    UNASSIGNED: Overall, in comparison to previous techniques, our methodology has shown much better results on k-Nearest Neighbors (KNN) by yielding the maximum precision, recall, F1, and F2 score values (92.00, 92.66, 92.33, and 92.52%, respectively) for AFD and (95.00% for each variable) for LFW dataset, which were used as training and testing datasets. The FRAI model may be potentially used in healthcare and criminology as well as many other applications where it is important to quickly identify face features such as fingerprint for a specific identification target.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    深度学习在识别复杂结构方面发挥着重要作用,与传统算法相比,它在训练和分类任务方面的表现优于传统算法。在这项工作中,开发了一种基于云的本地解决方案,用于将阿尔茨海默病(AD)分类为MRI扫描作为输入模态。多分类用于AD品种,分为四个阶段。为了利用预先训练的GoogLeNet模型的功能,采用迁移学习。GoogLeNet模型,它是为图像分类任务预先训练的,针对多类别AD分类的特定目的进行了微调。通过这个过程,达到了98%的更好的准确度。因此,使用GoogLeNet提出的体系结构开发了用于阿尔茨海默病预测的本地云Web应用程序。此应用程序使医生能够远程检查患者是否存在AD。
    Deep learning is playing a major role in identifying complicated structure, and it outperforms in term of training and classification tasks in comparison to traditional algorithms. In this work, a local cloud-based solution is developed for classification of Alzheimer\'s disease (AD) as MRI scans as input modality. The multi-classification is used for AD variety and is classified into four stages. In order to leverage the capabilities of the pre-trained GoogLeNet model, transfer learning is employed. The GoogLeNet model, which is pre-trained for image classification tasks, is fine-tuned for the specific purpose of multi-class AD classification. Through this process, a better accuracy of 98% is achieved. As a result, a local cloud web application for Alzheimer\'s prediction is developed using the proposed architectures of GoogLeNet. This application enables doctors to remotely check for the presence of AD in patients.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    基于卷积神经网络(CNN)的辣椒叶病识别是有趣的研究领域之一。然而,大多数现有的基于CNN的辣椒叶病检测模型在准确性和计算性能方面都是次优的。特别是,由于在大型领域中进行叶片病害识别需要大量的计算和内存消耗,因此在嵌入式便携式设备上应用CNN具有挑战性。因此,本文介绍了一种基于GoogLeNet架构的增强型轻量级模型。初始步骤涉及压缩Inception结构以减少模型参数,导致识别速度显着提高。此外,该网络结合了空间金字塔池化结构,以无缝地整合局部和全局特征。随后,所提出的改进模型已经在9183张图像的真实数据集上进行了训练,含有6种辣椒病。交叉验证结果表明,模型准确率为97.87%,比基于Inception-V1和Inception-V3的GoogLeNet高出6%。该模型的内存需求仅为10.3MB,减少了52.31%-86.69%,与GoogLeNet相比。我们还将该模型与现有的基于CNN的模型进行了比较,包括AlexNet,ResNet-50和MobileNet-V2。结果表明,该模型的平均推理时间减少了61.49%,41.78%和23.81%,分别。结果表明,所提出的增强模型在精度和计算效率方面都能显著提高性能,这有可能提高辣椒种植业的生产力。
    Pepper leaf disease identification based on convolutional neural networks (CNNs) is one of the interesting research areas. However, most existing CNN-based pepper leaf disease detection models are suboptimal in terms of accuracy and computing performance. In particular, it is challenging to apply CNNs on embedded portable devices due to a large amount of computation and memory consumption for leaf disease recognition in large fields. Therefore, this paper introduces an enhanced lightweight model based on GoogLeNet architecture. The initial step involves compressing the Inception structure to reduce model parameters, leading to a remarkable enhancement in recognition speed. Furthermore, the network incorporates the spatial pyramid pooling structure to seamlessly integrate local and global features. Subsequently, the proposed improved model has been trained on the real dataset of 9183 images, containing 6 types of pepper diseases. The cross-validation results show that the model accuracy is 97.87%, which is 6% higher than that of GoogLeNet based on Inception-V1 and Inception-V3. The memory requirement of the model is only 10.3 MB, which is reduced by 52.31%-86.69%, comparing to GoogLeNet. We have also compared the model with the existing CNN-based models including AlexNet, ResNet-50 and MobileNet-V2. The result shows that the average inference time of the proposed model decreases by 61.49%, 41.78% and 23.81%, respectively. The results show that the proposed enhanced model can significantly improve performance in terms of accuracy and computing efficiency, which has potential to improve productivity in the pepper farming industry.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    由于2019年冠状病毒病(COVID-19)的快速传播,识别这种疾病,死亡率和恢复率的预测被认为是全球面临的重大挑战之一。本研究分析了COVID-19在世界各地传播的发生,并建议使用基于人工智能(AI)的深度学习算法来检测COVID19患者的阳性病例,使用真实世界数据集的死亡率和恢复率。最初,不需要的数据,如介词,链接,标签等。,使用一些预处理技术来移除。之后,术语频率反术语频率(TF-IDF)和词袋(BoW)技术用于从预处理的数据集中提取特征。然后,执行Mayfly优化(MO)算法以从特征集中挑选相关特征。最后,两个深度学习过程,ResNet模型和GoogleNet模型,进行杂交以实现预测过程。我们的系统检查了两种不同类型的公开文本数据集,以识别COVID-19疾病,并使用这些数据集预测死亡率和恢复率。有四个不同的数据集来分析性能,其中提出的方法达到97.56%的准确率,比线性回归(LR)和多项式朴素贝叶斯(MNB)高1.40%,比随机森林(RF)和随机梯度提升(SGB)高3.39%,比决策树(DT)和Bagging技术高5.32%。与现有的机器学习模型相比,仿真结果表明,提出的混合深度学习方法在冠状病毒识别和未来死亡率预测研究中具有一定的应用价值。
    Due the quick spread of coronavirus disease 2019 (COVID-19), identification of that disease, prediction of mortality rate and recovery rate are considered as one of the critical challenges in the whole world. The occurrence of COVID-19 dissemination beyond the world is analyzed in this research and an artificial-intelligence (AI) based deep learning algorithm is suggested to detect positive cases of COVID19 patients, mortality rate and recovery rate using real-world datasets. Initially, the unwanted data like prepositions, links, hashtags etc., are removed using some pre-processing techniques. After that, term frequency inverse-term frequency (TF-IDF) andBag of Words (BoW) techniques are utilized to extract the features from pre-processed dataset. Then, Mayfly Optimization (MO) algorithm is performed to pick the relevant features from the set of features. Finally, two deep learning procedures, ResNet model and GoogleNet model, are hybridized to achieve the prediction process. Our system examines two different kinds of publicly available text datasets to identify COVID-19 disease as well as to predict mortality rate and recovery rate using those datasets. There are four different datasets are taken to analyse the performance, in which the proposed method achieves 97.56% accuracy which is 1.40% greater than Linear Regression (LR) and Multinomial Naive Bayesian (MNB), 3.39% higher than Random Forest (RF) and Stochastic gradient boosting (SGB) as well as 5.32% higher than Decision tree (DT) and Bagging techniques if first dataset. When compared to existing machine learning models, the simulation result indicates that a proposed hybrid deep learning method is valuable in corona virus identification and future mortality forecast study.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    视频胶囊内窥镜(VCE)是一种新兴技术,其允许以最小的侵入检查整个胃肠(GI)道。虽然传统的内镜活检程序是诊断大多数胃肠道疾病的金标准,它们受限于范围可以在管道中推进多远,并且也是侵入性的。VCE允许胃肠病学家通过可视化胃肠道的所有部分来详细研究胃肠道异常。它捕获连续的实时图像,因为它是由肠道运动推进胃肠道。即使VCE允许彻底检查,审查和分析长达8个小时的图像(编译为视频)是繁琐的,并不具有成本效益。为了为基于VCE的胃肠道疾病诊断的自动化铺平道路,检测胶囊的位置将允许更集中的分析以及在胃肠道的每个区域中的异常检测。在本文中,我们比较了4种深度卷积神经网络模型,用于特征提取和检测VCE图像捕获的胃肠道内的解剖部分。我们的结果表明,VGG-Net具有卓越的性能和最高的平均精度,精度,召回和,与其他最先进的架构相比,F1分数:GoogLeNet,AlexNet和,ResNet。
    Video capsule endoscope (VCE) is an emerging technology that allows examination of the entire gastrointestinal (GI) tract with minimal invasion. While traditional endoscopy with biopsy procedures are the gold standard for diagnosis of most GI diseases, they are limited by how far the scope can be advanced in the tract and are also invasive. VCE allows gastroenterologists to investigate GI tract abnormalities in detail with visualization of all parts of the GI tract. It captures continuous real time images as it is propelled in the GI tract by gut motility. Even though VCE allows for thorough examination, reviewing and analyzing up to eight hours of images (compiled as videos) is tedious and not cost effective. In order to pave way for automation of VCE-based GI disease diagnosis, detecting the location of the capsule would allow for a more focused analysis as well as abnormality detection in each region of the GI tract. In this paper, we compared four deep Convolutional Neural Network models for feature extraction and detection of the anatomical part within the GI tract captured by VCE images. Our results showed that VGG-Net has superior performance with the highest average accuracy, precision, recall and, F1-score compared to other state of the art architectures: GoogLeNet, AlexNet and, ResNet.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    道路状况评价是碎石路养护的关键部分。评估参数之一是松散砾石的数量,因为这决定了驾驶质量和安全性。松散的碎石会导致轮胎打滑,驾驶员失去控制。专家通过观察图像和笔记主观地评估路况。这种方法是劳动密集型的,并且容易判断错误;因此,其可靠性值得怀疑。道路管理机构寻找自动化和客观的测量系统。在这项研究中,使用了砾石撞击汽车底部的声学数据。评估了声学与碎石路上松散砾石状况之间的联系。应用传统的监督学习算法和卷积神经网络(CNN),并比较了它们的性能,以进行松散砾石声学分类。使用预训练的CNN的优点是它选择相关特征进行训练。此外,预先训练的网络具有不需要几天的训练或庞大的训练数据的优势。在监督学习中,集合袋式树算法对砾石和非砾石声音分类的准确率为97.5%,然而,在深度学习的情况下,预训练网络GoogLeNet的准确率为97.91%,用于对砾石声音的频谱图图像进行分类。
    Road condition evaluation is a critical part of gravel road maintenance. One of the assessed parameters is the amount of loose gravel, as this determines the driving quality and safety. Loose gravel can cause tires to slip and the driver to lose control. An expert assesses the road conditions subjectively by looking at images and notes. This method is labor-intensive and subject to error in judgment; therefore, its reliability is questionable. Road management agencies look for automated and objective measurement systems. In this study, acoustic data on gravel hitting the bottom of a car was used. The connection between the acoustics and the condition of loose gravel on gravel roads was assessed. Traditional supervised learning algorithms and convolution neural network (CNN) were applied, and their performances are compared for the classification of loose gravel acoustics. The advantage of using a pre-trained CNN is that it selects relevant features for training. In addition, pre-trained networks offer the advantage of not requiring days of training or colossal training data. In supervised learning, the accuracy of the ensemble bagged tree algorithm for gravel and non-gravel sound classification was found to be 97.5%, whereas, in the case of deep learning, pre-trained network GoogLeNet accuracy was 97.91% for classifying spectrogram images of the gravel sounds.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Sci-hub)

       PDF(Pubmed)

  • 文章类型: Journal Article
    The ongoing intense development of short-range radar systems and their improved capability of measuring small movements make these systems reliable solutions for the extraction of human vital signs in a contactless fashion. The continuous contactless monitoring of vital signs can be considered in a wide range of applications, such as remote healthcare solutions and context-aware smart sensor development. Currently, the provision of radar-recorded datasets of human vital signs is still an open issue. In this paper, we present a new frequency-modulated continuous wave (FMCW) radar-recorded vital sign dataset for 50 children aged less than 13 years. A clinically approved vital sign monitoring sensor was also deployed as a reference, and data from both sensors were time-synchronized. With the presented dataset, a new child age-group classification system based on GoogLeNet is proposed to develop a child safety sensor for smart vehicles. The radar-recorded vital signs of children are divided into several age groups, and the GoogLeNet framework is trained to predict the age of unknown human test subjects.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Sci-hub)

       PDF(Pubmed)

公众号