GoogleNet

GoogleNet
  • 文章类型: Journal Article
    PIN二极管,由于其结构简单,在高频大功率激励下具有可变电阻特性,通常在雷达前端用作限制器,以过滤高功率微波(HPM),以防止其电源进入内部电路并造成损坏。本文对PIN二极管的HPM效应进行了理论推导和研究,然后用优化的神经网络算法代替传统的物理建模来计算和预测PIN二极管限幅器的两类HPM限幅指标。我们针对以下两种预测场景中的每一种提出了神经网络模型:在不同HPM辐照下的时间-结温曲线的场景中,来自测试数据集的预测值与模拟值之间的加权均方误差(MSE)低于0.004。在预测PIN限制器的功率限制阈值时,插入损耗,以及不同HPM辐照下的最大隔离度,测试集预测值和模拟值的MSE均小于0.03。本研究提出的方法,应用优化的神经网络算法代替传统的物理建模算法来研究PIN二极管限幅器的高功率微波效应,显著提高了计算和仿真速度,降低了计算成本,为研究PIN二极管限幅器的高功率微波效应提供了一种新的方法。
    PIN diodes, due to their simple structure and variable resistance characteristics under high-frequency high-power excitation, are often used in radar front-end as limiters to filter high power microwaves (HPM) to prevent its power from entering the internal circuit and causing damage. This paper carries out theoretical derivation and research on the HPM effects of PIN diodes, and then uses an optimized neural network algorithm to replace traditional physical modeling to calculate and predict two types of HPM limiting indicators of PIN diode limiters. We proposes a neural network model for each of the following two prediction scenarios: in the scenario of time-junction temperature curves under different HPM irradiation, the weighted mean squared error (MSE) between the predicted values from the test dataset and the simulated values is below 0.004. While in predicting PIN limiter\'s power limitation threshold, insertion loss, and maximum isolation under different HPM irradiation, the MSE of the test set prediction values and simulation values are all less than 0.03. The method proposed in this research, which applies an optimized neural network algorithm to replace traditional physical modeling algorithms for studying the high-power microwave effects of PIN diode limiters, significantly improves the computational and simulation speed, reduces the calculation cost, and provides a new method for studying the high-power microwave effects of PIN diode limiters.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    多模式机器人音乐表演艺术领域由于其创新潜力而引起了极大的兴趣。传统的机器人在理解音乐表演中的情感和艺术表达方面面临局限性。因此,本文探讨了融合视觉和听觉感知的多模态机器人在音乐表演中的应用,以提高音乐表演的质量和艺术表现力。我们的方法涉及集成GRU(门控递归单元)和GoogLeNet模型进行情感分析。GRU模型处理音频数据并捕获音乐元素的时间动态,包括长期依赖,提取情感信息。GoogLeNet模型擅长图像处理,提取复杂的视觉细节和审美特征。这种协同作用加深了对音乐和视觉元素的理解,旨在产生更多情感共鸣和互动的机器人表演。实验结果证明了我们方法的有效性,显示出多模态机器人在音乐表现方面的显著改善。这些机器人,配备了我们的方法,提供高质量,有效唤起观众情感参与的艺术表演。在音乐表演中融合视听感知的多模态机器人丰富了艺术形式,并提供了多样化的人机交互。这项研究证明了多模式机器人在音乐表演中的潜力,促进技术与艺术的融合。它开辟了表演艺术和人机交互的新领域,提供独特和创新的体验。我们的发现为表演艺术领域多模式机器人的发展提供了宝贵的见解。
    The field of multimodal robotic musical performing arts has garnered significant interest due to its innovative potential. Conventional robots face limitations in understanding emotions and artistic expression in musical performances. Therefore, this paper explores the application of multimodal robots that integrate visual and auditory perception to enhance the quality and artistic expression in music performance. Our approach involves integrating GRU (Gated Recurrent Unit) and GoogLeNet models for sentiment analysis. The GRU model processes audio data and captures the temporal dynamics of musical elements, including long-term dependencies, to extract emotional information. The GoogLeNet model excels in image processing, extracting complex visual details and aesthetic features. This synergy deepens the understanding of musical and visual elements, aiming to produce more emotionally resonant and interactive robot performances. Experimental results demonstrate the effectiveness of our approach, showing significant improvements in music performance by multimodal robots. These robots, equipped with our method, deliver high-quality, artistic performances that effectively evoke emotional engagement from the audience. Multimodal robots that merge audio-visual perception in music performance enrich the art form and offer diverse human-machine interactions. This research demonstrates the potential of multimodal robots in music performance, promoting the integration of technology and art. It opens new realms in performing arts and human-robot interactions, offering a unique and innovative experience. Our findings provide valuable insights for the development of multimodal robots in the performing arts sector.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    基于卷积神经网络(CNN)的辣椒叶病识别是有趣的研究领域之一。然而,大多数现有的基于CNN的辣椒叶病检测模型在准确性和计算性能方面都是次优的。特别是,由于在大型领域中进行叶片病害识别需要大量的计算和内存消耗,因此在嵌入式便携式设备上应用CNN具有挑战性。因此,本文介绍了一种基于GoogLeNet架构的增强型轻量级模型。初始步骤涉及压缩Inception结构以减少模型参数,导致识别速度显着提高。此外,该网络结合了空间金字塔池化结构,以无缝地整合局部和全局特征。随后,所提出的改进模型已经在9183张图像的真实数据集上进行了训练,含有6种辣椒病。交叉验证结果表明,模型准确率为97.87%,比基于Inception-V1和Inception-V3的GoogLeNet高出6%。该模型的内存需求仅为10.3MB,减少了52.31%-86.69%,与GoogLeNet相比。我们还将该模型与现有的基于CNN的模型进行了比较,包括AlexNet,ResNet-50和MobileNet-V2。结果表明,该模型的平均推理时间减少了61.49%,41.78%和23.81%,分别。结果表明,所提出的增强模型在精度和计算效率方面都能显著提高性能,这有可能提高辣椒种植业的生产力。
    Pepper leaf disease identification based on convolutional neural networks (CNNs) is one of the interesting research areas. However, most existing CNN-based pepper leaf disease detection models are suboptimal in terms of accuracy and computing performance. In particular, it is challenging to apply CNNs on embedded portable devices due to a large amount of computation and memory consumption for leaf disease recognition in large fields. Therefore, this paper introduces an enhanced lightweight model based on GoogLeNet architecture. The initial step involves compressing the Inception structure to reduce model parameters, leading to a remarkable enhancement in recognition speed. Furthermore, the network incorporates the spatial pyramid pooling structure to seamlessly integrate local and global features. Subsequently, the proposed improved model has been trained on the real dataset of 9183 images, containing 6 types of pepper diseases. The cross-validation results show that the model accuracy is 97.87%, which is 6% higher than that of GoogLeNet based on Inception-V1 and Inception-V3. The memory requirement of the model is only 10.3 MB, which is reduced by 52.31%-86.69%, comparing to GoogLeNet. We have also compared the model with the existing CNN-based models including AlexNet, ResNet-50 and MobileNet-V2. The result shows that the average inference time of the proposed model decreases by 61.49%, 41.78% and 23.81%, respectively. The results show that the proposed enhanced model can significantly improve performance in terms of accuracy and computing efficiency, which has potential to improve productivity in the pepper farming industry.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    解决煤矿瓦斯和煤尘爆炸监测手段落后的问题,迟报,和低泄漏率,提出了一种基于GoogLeNet的煤矿瓦斯煤尘爆炸声音识别方法。在煤矿的重点监控区域安装采矿拾音器以收集工作设备和环境的声音后,用连续小波对采集到的声音进行分析,得到其尺度系数图。然后将其导入GoogLeNet,获得煤矿瓦斯和煤尘爆炸的识别模型。通过连续小波分析获得测试声音,获得比例系数图,引入到完成的训练识别模型中,以获得声音信号类,并通过实验验证。首先,连续小波分析从声音信号中提取的尺度系数图表明,瓦斯爆炸声和煤尘爆炸声的小波系数图的主客观指标相似度较高,但是这些声音和其他煤矿声音之间的区别更明显,有助于有效区分气体和煤尘爆炸声与其他声音。其次,可以获得GoogLeNet参数的实验结果。当dropout参数为0.5,初始学习率为0.001时,GoogLeNet建立的模型识别效果最优。根据选择的参数,训练损失,测试损失,训练识别率,测试模型的识别率均符合预期。最后,实验识别结果表明,该方法的识别率为97.38%,召回率为86.1%,并且对于测试数据与训练数据的比率为9:1的情况,准确率为100%,提出的GoogLeNet的整体识别效果明显优于vgg和Alexnet,能有效解决煤矿瓦斯和煤尘爆炸声音的欠采样问题,满足煤矿瓦斯和粉尘爆炸智能识别的需要。
    To solve the problems of backward means of coal mine gas and coal dust explosion monitoring, late reporting, and low leakage rate, a sound recognition method of coal mine gas and coal dust explosion based on GoogLeNet was proposed. After installing mining pickups in key monitoring areas of coal mines to collect the sounds of the working equipment and the environment, the collected sound was analyzed by continuous wavelet to obtain its scale coefficient map. This was then imported into GoogLeNet to obtain the recognition model of coal mine gas and coal dust explosions. The test sound was obtained by continuous wavelet analysis to obtain the scale coefficient map, brought into the completed training recognition model to obtain the sound signal class, and verified by experiment. Firstly, the scale coefficient map extracted from the sound signal by continuous wavelet analysis showed that the similarity between the subjective and objective indicators of the wavelet coefficient maps of the gas explosion sound and coal dust explosion sound was higher, but the difference between these and the rest of the coal mine sounds was clearer, helping to effectively distinguish gas and coal dust explosion sounds from other sounds. Secondly, the experimental results of GoogLeNet parameters can be obtained. When the dropout parameter is 0.5 and the initial learning rate is 0.001, the recognition effect of the model established by GoogLeNet was optimal. According to the selected parameters, the training loss, testing loss, training recognition rate, and testing recognition rate of the model are all in line with expectations. Finally, the experimental recognition results show that the recognition rate of the proposed method is 97.38%, the recall rate is 86.1%, and the accuracy rate is 100% for the case of a 9:1 ratio of test data to training data, and the overall recognition effect of the proposed GoogLeNet is significantly better than that of vgg and Alexnet, which can effectively solve the problem of under-sampling of coal mine gas and coal dust explosion sounds and can meet the need for the intelligent recognition of coal mine gas and dust explosions.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    背景:在三维医学图像中自动识别人体部位在许多临床应用中都很重要。然而,先前研究中提出的方法主要对每个二维(2D)切片进行独立分类,而不是将一批连续切片识别为特定的身体部位。
    目的:在本研究中,我们的目标是开发一种基于深度学习的方法,该方法旨在自动将计算机断层扫描(CT)和磁共振成像(MRI)扫描分为五个连续的身体部位:头部,脖子,胸部,腹部,还有骨盆.
    方法:开发了一个深度学习框架,分两个阶段识别身体部位。在第一个预分类阶段,将使用GoogLeNetInceptionv3架构的卷积神经网络(CNN)和长短期记忆(LSTM)网络组合起来对每个2D切片进行分类;CNN从单个切片中提取信息,而LSTM在连续切片中采用了丰富的上下文信息。在第二个后处理阶段,根据第一阶段的切片分类结果,通过确定它们之间的最佳边界,将输入扫描进一步划分为连续的身体部位。为了评估所提出方法的性能,使用662次CT和1434次MRI扫描。
    结果:与最先进的方法相比,我们的方法在2D切片分类中取得了非常好的性能,CT和MRI扫描的总体分类准确率为97.3%和98.2%,分别。此外,我们的方法进一步将整个扫描分为连续的身体部位,CT和MRI数据的平均边界误差为8.9和3.5mm,分别。
    结论:与最先进的方法相比,所提出的方法显着提高了切片分类的准确性,并根据切片分类结果进一步准确地将CT和MRI扫描分为连续的身体部位。所开发的方法可以用作各种计算机辅助诊断和医学图像分析方案中的重要步骤。
    BACKGROUND: The automatic recognition of human body parts in three-dimensional medical images is important in many clinical applications. However, methods presented in prior studies have mainly classified each two-dimensional (2D) slice independently rather than recognizing a batch of consecutive slices as a specific body part.
    OBJECTIVE: In this study, we aim to develop a deep learning-based method designed to automatically divide computed tomography (CT) and magnetic resonance imaging (MRI) scans into five consecutive body parts: head, neck, chest, abdomen, and pelvis.
    METHODS: A deep learning framework was developed to recognize body parts in two stages. In the first preclassification stage, a convolutional neural network (CNN) using the GoogLeNet Inception v3 architecture and a long short-term memory (LSTM) network were combined to classify each 2D slice; the CNN extracted information from a single slice, whereas the LSTM employed rich contextual information among consecutive slices. In the second postprocessing stage, the input scan was further partitioned into consecutive body parts by identifying the optimal boundaries between them based on the slice classification results of the first stage. To evaluate the performance of the proposed method, 662 CT and 1434 MRI scans were used.
    RESULTS: Our method achieved a very good performance in 2D slice classification compared with state-of-the-art methods, with overall classification accuracies of 97.3% and 98.2% for CT and MRI scans, respectively. Moreover, our method further divided whole scans into consecutive body parts with mean boundary errors of 8.9 and 3.5 mm for CT and MRI data, respectively.
    CONCLUSIONS: The proposed method significantly improved the slice classification accuracy compared with state-of-the-art methods, and further accurately divided CT and MRI scans into consecutive body parts based on the results of slice classification. The developed method can be employed as an important step in various computer-aided diagnosis and medical image analysis schemes.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Letter
    暂无摘要。
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    Identification of Chinese medicinal materials is a fundamental part and an important premise of the modern Chinese medicinal materials industry. As for the traditional Chinese medicinal materials that imitate wild cultivation, due to their scattered, irregular, and fine-grained planting characteristics, the fine classification using traditional classification methods is not accurate. Therefore, a deep convolution neural network model is used for imitating wild planting. Identification of Chinese herbal medicines. This study takes Lonicera japonica remote sensing recognition as an example, and proposes a method for fine classification of L. japonica based on a deep convolutional neural network model. The GoogLeNet network model is used to learn a large number of training samples to extract L. japonica characteristics from drone remote sensing images. Parameters, further optimize the network structure, and obtain a L. japonica recognition model. The research results show that the deep convolutional neural network based on GoogLeNet can effectively extract the L. japonica information that is relatively fragmented in the image, and realize the fine classification of L. japonica. After training and optimization, the overall classification accuracy of L. japonica can reach 97.5%, and total area accuracy is 94.6%, which can provide a reference for the application of deep convolutional neural network method in remote sensing classification of Chinese medicinal materials.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    OBJECTIVE: Convolutional neural networks (CNNs) play an important role in the field of medical image segmentation. Among many kinds of CNNs, the U-net architecture is one of the most famous fully convolutional network architectures for medical semantic segmentation tasks. Recent work shows that the U-net network can be substantially deeper thus resulting in improved performance on segmentation tasks. Though adding more layers directly into network is a popular way to make a network deeper, it may lead to gradient vanishing or redundant computation during training.
    METHODS: A novel CNN architecture is proposed that integrates the Inception-Res module and densely connecting convolutional module into the U-net architecture. The proposed network model consists of the following parts: firstly, the Inception-Res block is designed to increase the width of the network by replacing the standard convolutional layers; secondly, the Dense-Inception block is designed to extract features and make the network more deep without additional parameters; thirdly, the down-sampling block is adopted to reduce the size of feature maps to accelerate learning and the up-sampling block is used to resize the feature maps.
    RESULTS: The proposed model is tested on images of blood vessel segmentations from retina images, the lung segmentation of CT Data from the benchmark Kaggle datasets and the MRI scan brain tumor segmentation datasets from MICCAI BraTS 2017. The experimental results show that the proposed method can provide better performance on these two tasks compared with the state-of-the-art algorithms. The results reach an average Dice score of 0.9857 in the lung segmentation. For the blood vessel segmentation, the results reach an average Dice score of 0.9582. For the brain tumor segmentation, the results reach an average Dice score of 0.9867.
    CONCLUSIONS: The experiments highlighted that combining the inception module with dense connections in the U-Net architecture is a promising approach for semantic medical image segmentation.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Sci-hub)

  • 文章类型: Journal Article
    随着医学影像学理论和技术的发展,大多数肿瘤都可以在早期发现。然而,卵巢囊肿的性质缺乏准确的判断,导致许多良性结节患者仍然需要细针穿刺(FNA)活检或手术,增加患者的身体痛苦和精神压力以及不必要的医疗费用。因此,我们提出了一种图像诊断系统,用于在彩色超声图像中对卵巢囊肿进行分类,新应用由深度学习网络的高级特征和纹理描述符的低级特征融合的图像特征。首先,增强超声图像以提高训练数据集的质量,并且从每个图像中提取旋转不变均匀局部二值模式(ULBP)特征作为低级纹理特征。然后,通过微调的GoogLeNet神经网络提取的高级深度特征和低级ULBP特征被归一化并级联为一个融合特征,该融合特征可以表示语义上下文和分布在图像中的纹理模式。最后,将融合特征输入到成本敏感随机森林分类器以将图像分类为“恶性”和“良性”。深度神经网络从医学超声图像中提取的高级特征能够反映病变区域的视觉特征,而低级纹理特征可以描述边缘,强度的方向和分布。实验结果表明,两类特征的组合可以描述病变区域与其他区域的差异,以及恶性和良性卵巢囊肿病变区域之间的差异。
    With the development of theories and technologies in medical imaging, most of the tumors can be detected in the early stage. However, the nature of ovarian cysts lacks accurate judgement, leading to that many patients with benign nodules still need Fine Needle Aspiration (FNA) biopsies or surgeries, increasing the physical pain and mental pressure of patients as well as unnecessary medical health care costs. Therefore, we present an image diagnosis system for classifying the ovarian cysts in color ultrasound images, which novelly applies the image features fused by both high-level features from deep learning network and low-level features from texture descriptor. Firstly, the ultrasound images are enhanced to improve the quality of training data set and the rotation invariant uniform local binary pattern (ULBP) features are extracted from each of the images as the low-level texture features. Then the high-level deep features extracted by the fine-tuned GoogLeNet neural network and the low-level ULBP features are normalized and cascaded as one fusion feature that can represent both the semantic context and the texture patterns distributed in the image. Finally, the fusion features are input to the Cost-sensitive Random Forest classifier to classify the images into \"malignant\" and \"benign\". The high-level features extracted by the deep neural network from the medical ultrasound image can reflect the visual features of the lesion region, while the low-level texture features can describe the edges, direction and distribution of intensities. Experimental results indicate that the combination of the two types of features can describe the differences between the lesion regions and other regions, and the differences between lesions regions of malignant and benign ovarian cysts.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    In recent years, increasing human data comes from image sensors. In this paper, a novel approach combining convolutional pose machines (CPMs) with GoogLeNet is proposed for human pose estimation using image sensor data. The first stage of the CPMs directly generates a response map of each human skeleton\'s key points from images, in which we introduce some layers from the GoogLeNet. On the one hand, the improved model uses deeper network layers and more complex network structures to enhance the ability of low level feature extraction. On the other hand, the improved model applies a fine-tuning strategy, which benefits the estimation accuracy. Moreover, we introduce the inception structure to greatly reduce parameters of the model, which reduces the convergence time significantly. Extensive experiments on several datasets show that the improved model outperforms most mainstream models in accuracy and training time. The prediction efficiency of the improved model is improved by 1.023 times compared with the CPMs. At the same time, the training time of the improved model is reduced 3.414 times. This paper presents a new idea for future research.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Sci-hub)

       PDF(Pubmed)

公众号