lightweight

轻量级
  • 文章类型: Journal Article
    在错综复杂的环境中有效而准确地识别西红柿对于推进番茄收获的自动化至关重要。当前的对象检测算法速度慢,对遮挡和小番茄的识别准确率低。
    为了增强在复杂环境中对西红柿的检测,提出了一种轻量级温室番茄目标检测模型S-YOLO,基于YOLOv8s的几个关键改进:(1)为YOLOv8s量身定制的轻量级GSConv_SlimNeck结构是创新的,显着减少模型参数,以优化模型颈部以进行轻量化模型采集。(2)设计了一种改进的α-SimSPPF结构,有效提高了番茄的检测精度。(3)提出了一种增强型β-SIoU算法,优化了训练过程,提高了重叠番茄识别的准确率。(4)集成了SE注意模块,以使模型能够捕获更具代表性的温室番茄特征,从而提高检测精度。
    实验结果表明,增强的S-YOLO模型显著提高了检测精度,实现轻量化模型设计,并表现出快速的检测速度。实验结果表明,S-YOLO模型显著提高了检测精度,达到96.60%的精度,平均精度(mAP)92.46%,和74.05FPS的检测速度,提高了5.25%,2.1%,和3.49FPS分别比原来的模型。模型参数仅为9.11M,S-YOLO优于CenterNet等模型,YOLOv3,YOLOv4,YOLOv5m,YOLOv7和YOLOv8,有效地解决了遮挡和小番茄的识别准确率低的问题。
    S-YOLO模型的轻巧特性使其适用于番茄采摘机器人的视觉系统,为基于移动边缘计算的设施环境下的机器人目标识别和收获操作提供技术支持。
    UNASSIGNED: Efficiently and precisely identifying tomatoes amidst intricate surroundings is essential for advancing the automation of tomato harvesting. Current object detection algorithms are slow and have low recognition accuracy for occluded and small tomatoes.
    UNASSIGNED: To enhance the detection of tomatoes in complex environments, a lightweight greenhouse tomato object detection model named S-YOLO is proposed, based on YOLOv8s with several key improvements: (1) A lightweight GSConv_SlimNeck structure tailored for YOLOv8s was innovatively constructed, significantly reducing model parameters to optimize the model neck for lightweight model acquisition. (2) An improved version of the α-SimSPPF structure was designed, effectively enhancing the detection accuracy of tomatoes. (3) An enhanced version of the β-SIoU algorithm was proposed to optimize the training process and improve the accuracy of overlapping tomato recognition. (4) The SE attention module is integrated to enable the model to capture more representative greenhouse tomato features, thereby enhancing detection accuracy.
    UNASSIGNED: Experimental results demonstrate that the enhanced S-YOLO model significantly improves detection accuracy, achieves lightweight model design, and exhibits fast detection speeds. Experimental results demonstrate that the S-YOLO model significantly enhances detection accuracy, achieving 96.60% accuracy, 92.46% average precision (mAP), and a detection speed of 74.05 FPS, which are improvements of 5.25%, 2.1%, and 3.49 FPS respectively over the original model. With model parameters at only 9.11M, the S-YOLO outperforms models such as CenterNet, YOLOv3, YOLOv4, YOLOv5m, YOLOv7, and YOLOv8s, effectively addressing the low recognition accuracy of occluded and small tomatoes.
    UNASSIGNED: The lightweight characteristics of the S-YOLO model make it suitable for the visual system of tomato-picking robots, providing technical support for robot target recognition and harvesting operations in facility environments based on mobile edge computing.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    红茶是中国第二常见的茶。发酵是其生产中最关键的过程之一,影响成品的质量,无论是不足还是过度。目前,红茶发酵程度的测定完全依靠人工经验。这导致红茶的质量不一致。为了解决这个问题,我们使用机器视觉技术根据图像来区分红茶的发酵程度,本文提出了一种结合知识蒸馏的轻量级卷积神经网络(CNN)来判别红茶的发酵程度。在比较了12种CNN模型后,考虑到模型的大小和歧视的表现,以及教师模型的选择原则,选择Shufflenet_v2_x1.0作为学生模型,并选择Efficientnet_v2作为教师模型。然后,熵损失被焦点损失取代。最后,对于0.6、0.7、0.8、0.9的蒸馏损失率,软目标知识蒸馏(ST),掩蔽生成蒸馏(MGD),保持相似度的知识蒸馏(SPKD),和注意力转移(AT)四种知识蒸馏方法在Shufflenet_v2_x1.0模型中提取知识的性能进行了测试。结果表明,当蒸馏损失率为0.8,采用MGD方法时,蒸馏后的模型判别性能最好。这种设置有效地提高了辨别性能,而不会增加参数的数量和计算量。模型的P,R和F1值分别达到0.9208、0.9190和0.9192。实现了对红茶发酵程度的精确判别。这满足了客观红茶发酵判断的要求,为红茶的智能化加工提供了技术支持。
    Black tea is the second most common type of tea in China. Fermentation is one of the most critical processes in its production, and it affects the quality of the finished product, whether it is insufficient or excessive. At present, the determination of black tea fermentation degree completely relies on artificial experience. It leads to inconsistent quality of black tea. To solve this problem, we use machine vision technology to distinguish the degree of fermentation of black tea based on images, this paper proposes a lightweight convolutional neural network (CNN) combined with knowledge distillation to discriminate the degree of fermentation of black tea. After comparing 12 kinds of CNN models, taking into account the size of the model and the performance of discrimination, as well as the selection principle of teacher models, Shufflenet_v2_x1.0 is selected as the student model, and Efficientnet_v2 is selected as the teacher model. Then, CrossEntropy Loss is replaced by Focal Loss. Finally, for Distillation Loss ratios of 0.6, 0.7, 0.8, 0.9, Soft Target Knowledge Distillation (ST), Masked Generative Distillation (MGD), Similarity-Preserving Knowledge Distillation (SPKD), and Attention Transfer (AT) four knowledge distillation methods are tested for their performance in distilling knowledge from the Shufflenet_v2_x1.0 model. The results show that the model discrimination performance after distillation is the best when the Distillation Loss ratio is 0.8 and the MGD method is used. This setup effectively improves the discrimination performance without increasing the number of parameters and computation volume. The model\'s P, R and F1 values reach 0.9208, 0.9190 and 0.9192, respectively. It achieves precise discrimination of the fermentation degree of black tea. This meets the requirements of objective black tea fermentation judgment and provides technical support for the intelligent processing of black tea.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    玉米叶片数是评估植物生长和调节种群结构的重要指标。然而,传统的叶片计数方法主要依靠人工,这既耗时又紧张,而现有的图像处理方法精度低,适应性差,难以满足实际应用的标准。为了准确检测玉米的生长状况,本研究提出了一种改进的轻质YOLOv8玉米叶片检测和计数方法。首先,用StarNet网络取代YOLOv8网络的骨干,并引入卷积和注意力融合模块(CAFM),结合局部卷积和全局注意力机制,增强不同通道信息的特征表示和融合能力。其次,在颈部网络部分,StarBlock模块用于改进C2f模块,以捕获更复杂的特征,同时通过跳转连接保留原始特征信息,以提高训练稳定性和性能。最后,轻量级共享卷积检测头(LSCD)用于减少重复计算并提高计算效率。实验结果表明,召回,改进模型的mAP50为97.9%,95.5%,97.5%,模型参数和模型尺寸的数量为1.8M和3.8MB,与YOLOv8相比,分别减少了40.86%和39.68%。研究表明,该模型提高了玉米叶片检测的准确性,协助育种者做出科学决策,为玉米叶数移动端检测装置的部署和应用提供参考,为玉米生长质量评价提供技术支持。
    The number of maize leaves is an important indicator for assessing plant growth and regulating population structure. However, the traditional leaf counting method mainly relies on manual work, which is both time-consuming and straining, while the existing image processing methods have low accuracy and poor adaptability, making it difficult to meet the standards for practical application. To accurately detect the growth status of maize, an improved lightweight YOLOv8 maize leaf detection and counting method was proposed in this study. Firstly, the backbone of the YOLOv8 network is replaced using the StarNet network and the convolution and attention fusion module (CAFM) is introduced, which combines the local convolution and global attention mechanisms to enhance the ability of feature representation and fusion of information from different channels. Secondly, in the neck network part, the StarBlock module is used to improve the C2f module to capture more complex features while preserving the original feature information through jump connections to improve training stability and performance. Finally, a lightweight shared convolutional detection head (LSCD) is used to reduce repetitive computations and improve computational efficiency. The experimental results show that the precision, recall, and mAP50 of the improved model are 97.9%, 95.5%, and 97.5%, and the numbers of model parameters and model size are 1.8 M and 3.8 MB, which are reduced by 40.86% and 39.68% compared to YOLOv8. This study shows that the model improves the accuracy of maize leaf detection, assists breeders in making scientific decisions, provides a reference for the deployment and application of maize leaf number mobile end detection devices, and provides technical support for the high-quality assessment of maize growth.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    作为最重要的经济作物之一,葡萄因其高产而备受关注,丰富的营养价值,和各种健康益处。识别葡萄束对于保持葡萄的质量和数量至关重要,以及管理病虫害。近年来,自动化设备与物体检测技术的结合有助于实现这一目标。然而,现有的轻量级目标检测算法往往以牺牲检测精度为代价来换取处理速度,这可能会在实际应用中造成障碍。因此,本文提出了一种轻量级的检测方法YOLOv8s-葡萄,其中包含几个有效的改进点,包括改进的有效信道注意力(MECA),苗条的脖子,新的空间金字塔快速汇集(NSPPF),动态上采样器(DySample),和具有最小点距离(MPDIoU)的并集上的交点。在提出的方法中,MECA和NSPPF增强了骨干的特征提取能力,使它能够更好地捕获关键信息。细长的脖子减少了多余的功能,降低计算复杂度,并有效地重用浅层特征以获得更详细的信息,进一步提高检测精度。DySample实现了出色的性能,同时保持了较低的计算成本,从而证明了很高的实用性和快速检测能力。MPDIoU通过更快的收敛和更精确的回归结果来提高检测精度。实验结果表明,与其他方法相比,该方法在葡萄藤束检测数据集和葡萄藤束条件检测数据集中表现更好,与YOLOv8相比,平均精度(MAP50-95)分别提高了2.4%和2.6%,分别。同时,该方法的计算复杂度和参数也降低了,每秒减少2.3Giga浮点运算和150万个参数。因此,可以得出结论,所提出的方法,整合了这些改进,实现了轻量级和高精度的检测,证明其在识别葡萄束和评估生物物理异常方面的有效性。
    As one of the most important economic crops, grapes have attracted considerable attention due to their high yield, rich nutritional value, and various health benefits. Identifying grape bunches is crucial for maintaining the quality and quantity of grapes, as well as managing pests and diseases. In recent years, the combination of automated equipment with object detection technology has been instrumental in achieving this. However, existing lightweight object detection algorithms often sacrifice detection precision for processing speed, which may pose obstacles in practical applications. Therefore, this thesis proposes a lightweight detection method named YOLOv8s-grape, which incorporates several effective improvement points, including modified efficient channel attention (MECA), slim-neck, new spatial pyramid pooling fast (NSPPF), dynamic upsampler (DySample), and intersection over union with minimum point distance (MPDIoU). In the proposed method, MECA and NSPPF enhance the feature extraction capability of the backbone, enabling it to better capture crucial information. Slim-neck reduces redundant features, lowers computational complexity, and effectively reuses shallow features to obtain more detailed information, further improving detection precision. DySample achieves excellent performance while maintaining lower computational costs, thus demonstrating high practicality and rapid detection capability. MPDIoU enhances detection precision through faster convergence and more precise regression results. Experimental results show that compared to other methods, this approach performs better in the grapevine bunch detection dataset and grapevine bunch condition detection dataset, with mean average precision (mAP50-95) increasing by 2.4% and 2.6% compared to YOLOv8s, respectively. Meanwhile, the computational complexity and parameters of the method are also reduced, with a decrease of 2.3 Giga floating-point operations per second and 1.5 million parameters. Therefore, it can be concluded that the proposed method, which integrates these improvements, achieves lightweight and high-precision detection, demonstrating its effectiveness in identifying grape bunches and assessing biophysical anomalies.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    低抗弯强度和韧性对胶凝材料提出了持久的挑战。作为水泥的主要水化产物,水化硅酸钙(C-S-H)在胶凝材料的力学性能中起重要作用,这阻碍了机械增强。灵感来自天然珍珠质的“砖与砂浆”微观结构,本文提出了一种结合冷冻铸造的方法,冷冻干燥,原位聚合,并进行热压,以制造具有高抗弯强度的C-S-H珍珠质,高韧性,和轻量级。聚(丙烯酰胺-共-丙烯酸)用于分散C-S-H和增韧C-S-H结构单元,它的功能是“砖块”,而聚(甲基丙烯酸甲酯)浸渍为“砂浆”。抗弯强度,韧性,C-S-H珍珠母密度达到124兆帕,5173kJ/m3和0.98g/cm3。C-S-H珍珠质的抗弯强度和韧性分别是水泥浆的18倍和1230倍,分别,密度降低了60%,优于现有的胶凝材料和天然珍珠层。本研究建立了材料成分之间的关系,制造工艺,微观结构,和机械性能,促进高性能C-S-H基和水泥基复合材料的设计,用于可扩展的工程应用。
    Low flexural strength and toughness have posed enduring challenges to cementitious materials. As the main hydration product of cement, calcium silicate hydrate (C-S-H) plays important roles in the mechanical performance of cementitious materials while exhibiting random microstructures with pores and defects, which hinder mechanical enhancement. Inspired by the \"brick-and-mortar\" microstructure of natural nacre, this paper presents a method combining freeze casting, freeze-drying, in situ polymerization, and hot pressing to fabricate C-S-H nacre with high flexural strength, high toughness, and lightweight. Poly(acrylamide-co-acrylic acid) was used to disperse C-S-H and toughen C-S-H building blocks, which function as \"bricks\", while poly(methyl methacrylate) was impregnated as \"mortar\". The flexural strength, toughness, and density of C-S-H nacre reached 124 MPa, 5173 kJ/m3, and 0.98 g/cm3, respectively. The flexural strength and toughness of the C-S-H nacre are 18 and 1230 times higher than those of cement paste, respectively, with a 60% reduction in density, outperforming existing cementitious materials and natural nacre. This research establishes the relationship between material composition, fabrication process, microstructure, and mechanical performance, facilitating the design of high-performance C-S-H-based and cement-based composites for scalable engineering applications.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    在野外对红花的有效检测对于实现自动视觉导航和采集系统至关重要。由于红花簇的物理尺寸小,它们密集的空间分布,以及现场场景的复杂性,当前的目标检测技术在红花检测中面临着一些挑战,如精度不足和计算要求高。因此,介绍了一种改进的基于YOLOv5的红花目标检测模型——红花-YOLO(SF-YOLO)。该模型采用Ghost_conv代替骨干网络中的传统卷积块,显著提高计算效率。此外,将CBAM注意力机制集成到骨干网中,并引入了组合的LCIOUNWD损失函数,以允许更精确的特征提取,增强的自适应融合能力,加速损耗收敛。锚箱,通过K均值聚类更新,用来替换原来的锚,使模型能够更好地适应田间红花的多尺度信息。数据增强技术,如高斯模糊,噪声添加,锐化,和通道混洗被应用于数据集,以保持对光照变化的鲁棒性,噪音,和视觉角度。实验结果表明,SF-YOLO超越了原有的YOLOv5s模型,GFlops和Params从15.8G减少到13.2G,从7.013减少到5.34M,分别,分别下降16.6%和23.9%。同时,SF-YOLO的mAP0.5增加1.3%,达到95.3%。这项工作提高了红花在复杂农业环境中检测的准确性,为后续红花作业中的自主视觉导航和自动化无损采收技术提供参考。
    The effective detection of safflower in the field is crucial for implementing automated visual navigation and harvesting systems. Due to the small physical size of safflower clusters, their dense spatial distribution, and the complexity of field scenes, current target detection technologies face several challenges in safflower detection, such as insufficient accuracy and high computational demands. Therefore, this paper introduces an improved safflower target detection model based on YOLOv5, termed Safflower-YOLO (SF-YOLO). This model employs Ghost_conv to replace traditional convolution blocks in the backbone network, significantly enhancing computational efficiency. Furthermore, the CBAM attention mechanism is integrated into the backbone network, and a combined L C I O U + N W D loss function is introduced to allow for more precise feature extraction, enhanced adaptive fusion capabilities, and accelerated loss convergence. Anchor boxes, updated through K-means clustering, are used to replace the original anchors, enabling the model to better adapt to the multi-scale information of safflowers in the field. Data augmentation techniques such as Gaussian blur, noise addition, sharpening, and channel shuffling are applied to the dataset to maintain robustness against variations in lighting, noise, and visual angles. Experimental results demonstrate that SF-YOLO surpasses the original YOLOv5s model, with reductions in GFlops and Params from 15.8 to 13.2 G and 7.013 to 5.34 M, respectively, representing decreases of 16.6% and 23.9%. Concurrently, SF-YOLO\'s mAP0.5 increases by 1.3%, reaching 95.3%. This work enhances the accuracy of safflower detection in complex agricultural environments, providing a reference for subsequent autonomous visual navigation and automated non-destructive harvesting technologies in safflower operations.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    为了缩短检测时间,提高嵌入式设备的平均精度,提出了一种轻量级和高精度的模型来检测复杂环境中的百香果(例如,有背光,遮挡,重叠,sun,云,或下雨)。首先,用轻量级的GhostNet模型代替YOLOv5的骨干网络,减少了参数数量和计算复杂度,同时提高了检测速度。第二,在骨干网络中增加一个新的特征分支,并重建颈部网络中的特征融合层,以有效地结合低级和高级特征,这提高了模型的准确性,同时保持了其轻量级。最后,使用知识蒸馏方法将知识从能力较强的教师模型转移到能力较弱的学生模型,显著提高了检测精度。改进的模型表示为G-YOLO-NK。G-YOLO-NK网络的平均准确率为96.00%,比原来的YOLOv5s型号高出1.00%。此外,模型大小为7.14MB,原来型号的一半,在JetsonNano上实现时,其实时检测帧速率为11.25FPS。发现所提出的模型在平均精度和检测性能方面优于最先进的模型。本工作为复杂果园场景中百香果的实时检测提供了一种有效的模型,为果园采摘机器人的发展提供了宝贵的技术支撑,大大提高了果园的智能化水平。
    In order to shorten detection times and improve average precision in embedded devices, a lightweight and high-accuracy model is proposed to detect passion fruit in complex environments (e.g., with backlighting, occlusion, overlap, sun, cloud, or rain). First, replacing the backbone network of YOLOv5 with a lightweight GhostNet model reduces the number of parameters and computational complexity while improving the detection speed. Second, a new feature branch is added to the backbone network and the feature fusion layer in the neck network is reconstructed to effectively combine the lower- and higher-level features, which improves the accuracy of the model while maintaining its lightweight nature. Finally, a knowledge distillation method is used to transfer knowledge from the more capable teacher model to the less capable student model, significantly improving the detection accuracy. The improved model is denoted as G-YOLO-NK. The average accuracy of the G-YOLO-NK network is 96.00%, which is 1.00% higher than that of the original YOLOv5s model. Furthermore, the model size is 7.14 MB, half that of the original model, and its real-time detection frame rate is 11.25 FPS when implemented on the Jetson Nano. The proposed model is found to outperform state-of-the-art models in terms of average precision and detection performance. The present work provides an effective model for real-time detection of passion fruit in complex orchard scenes, offering valuable technical support for the development of orchard picking robots and greatly improving the intelligence level of orchards.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    为了应对复杂的城市街道环境中垃圾的准确识别和定位的挑战,本文提出了EcoDetect-YOLO,基于YOLOv5s框架的垃圾暴露检测算法,利用本研究构建的复杂环境废物暴露检测数据集。最初,卷积块注意模块(CBAM)被集成在特征金字塔网络的第二级(P2)和特征金字塔网络的第三级(P3)层之间,以优化相关垃圾特征的提取,同时减轻背景噪声。随后,P2小目标检测头增强了模型识别小垃圾目标的功效。最后,引入了双向特征金字塔网络(BiFPN),以增强模型的深度特征融合能力。实验结果表明EcoDetect-YOLO对城市环境的适应性和优越的小目标检测能力,有效识别九种垃圾,如纸和塑料垃圾。与基线YOLOv5s模型相比,EcoDetect-YOLO的mAP0.5增长了4.7%,达到58.1%,紧凑的型号尺寸为15.7MB,FPS为39.36。值得注意的是,即使有强烈的噪音,该模型保持了超过50%的mAP0.5,强调其稳健性。总之,EcoDetect-YOLO,正如本文所提出的,拥有高精度,效率,和紧凑,使其适合部署在移动设备上,以实时检测和管理城市垃圾暴露,从而推进城市自动化治理和数字经济发展。
    In response to the challenges of accurate identification and localization of garbage in intricate urban street environments, this paper proposes EcoDetect-YOLO, a garbage exposure detection algorithm based on the YOLOv5s framework, utilizing an intricate environment waste exposure detection dataset constructed in this study. Initially, a convolutional block attention module (CBAM) is integrated between the second level of the feature pyramid etwork (P2) and the third level of the feature pyramid network (P3) layers to optimize the extraction of relevant garbage features while mitigating background noise. Subsequently, a P2 small-target detection head enhances the model\'s efficacy in identifying small garbage targets. Lastly, a bidirectional feature pyramid network (BiFPN) is introduced to strengthen the model\'s capability for deep feature fusion. Experimental results demonstrate EcoDetect-YOLO\'s adaptability to urban environments and its superior small-target detection capabilities, effectively recognizing nine types of garbage, such as paper and plastic trash. Compared to the baseline YOLOv5s model, EcoDetect-YOLO achieved a 4.7% increase in mAP0.5, reaching 58.1%, with a compact model size of 15.7 MB and an FPS of 39.36. Notably, even in the presence of strong noise, the model maintained a mAP0.5 exceeding 50%, underscoring its robustness. In summary, EcoDetect-YOLO, as proposed in this paper, boasts high precision, efficiency, and compactness, rendering it suitable for deployment on mobile devices for real-time detection and management of urban garbage exposure, thereby advancing urban automation governance and digital economic development.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    在单幅图像超分辨率背景下,特征提取起着举足轻重的作用。尽管如此,依靠单一的特征提取方法往往会破坏特征表示的全部潜力,妨碍模型的整体性能。为了解决这个问题,这项研究介绍了广泛的活化特征蒸馏网络(WFDN),通过双路径学习实现单幅图像的超分辨率。最初,采用双路径并行网络结构,利用剩余网络作为骨干,并结合全局剩余连接,以增强功能开发并加快网络融合。随后,采用了特征蒸馏块,其特点是训练速度快,参数计数低。同时,整合了广泛的激活机制,以进一步提高高频特征的表示能力。最后,引入门控融合机制对双分支提取的特征信息进行加权融合。该机制增强了重建性能,同时减轻了信息冗余。大量的实验表明,与最先进的方法相比,该算法获得了稳定和优越的结果,对四个基准数据集进行的定量评估指标测试证明了这一点。此外,我们的WFDN擅长重建具有更丰富详细纹理的图像,更现实的线条,更清晰的结构,肯定了其非凡的优越性和稳健性。
    Feature extraction plays a pivotal role in the context of single image super-resolution. Nonetheless, relying on a single feature extraction method often undermines the full potential of feature representation, hampering the model\'s overall performance. To tackle this issue, this study introduces the wide-activation feature distillation network (WFDN), which realizes single image super-resolution through dual-path learning. Initially, a dual-path parallel network structure is employed, utilizing a residual network as the backbone and incorporating global residual connections to enhance feature exploitation and expedite network convergence. Subsequently, a feature distillation block is adopted, characterized by fast training speed and a low parameter count. Simultaneously, a wide-activation mechanism is integrated to further enhance the representational capacity of high-frequency features. Lastly, a gated fusion mechanism is introduced to weight the fusion of feature information extracted from the dual branches. This mechanism enhances reconstruction performance while mitigating information redundancy. Extensive experiments demonstrate that the proposed algorithm achieves stable and superior results compared to the state-of-the-art methods, as evidenced by quantitative evaluation metrics tests conducted on four benchmark datasets. Furthermore, our WFDN excels in reconstructing images with richer detailed textures, more realistic lines, and clearer structures, affirming its exceptional superiority and robustness.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    目前,利用AI图像识别技术对海量电网输电线路巡检图片进行智能缺陷检测是一种高效、流行的方法。通常,缺陷检测算法模型的构建有两条技术路线:一是使用轻量级网络,这提高了效率,但它通常只能针对少数类型的缺陷,并可能降低检测精度;另一种是使用复杂的网络模型,这提高了准确性,并且可以同时识别多种类型的缺陷,但是它的计算量大,效率低。为了保持模型的高检测精度以及其轻量化结构,提出了一种轻量高效的基于DCP-YOLOv8的输电线路多类型缺陷检测方法。该方法采用可变形卷积(C2f_DCNv3)来增强缺陷特征提取能力,并设计了一种重新参数化的交叉相位特征融合结构(RCSP),将高层语义特征与低层空间特征进行优化和融合,从而提高模型识别不同尺度缺陷的能力,同时显著降低模型参数;它结合了动态检测头和可变形卷积v3的检测头(DCNv3-Dyhead),以增强特征表达能力和上下文信息的利用率,进一步提高检测精度。实验结果表明,在一个包含20条真实输电线路缺陷的数据集上,该方法将平均准确度(mAP@0.5)提高到72.2%,增长4.3%,与最轻的基线YOLOv8n模型相比;模型参数数量仅为2.8M,减少9.15%,每秒处理帧数(FPS)达到103,满足实时检测需求。在多类型缺陷检测的场景中,它有效地平衡了检测精度和性能与定量泛化。
    Currently, the intelligent defect detection of massive grid transmission line inspection pictures using AI image recognition technology is an efficient and popular method. Usually, there are two technical routes for the construction of defect detection algorithm models: one is to use a lightweight network, which improves the efficiency, but it can generally only target a few types of defects and may reduce the detection accuracy; the other is to use a complex network model, which improves the accuracy, and can identify multiple types of defects at the same time, but it has a large computational volume and low efficiency. To maintain the model\'s high detection accuracy as well as its lightweight structure, this paper proposes a lightweight and efficient multi type defect detection method for transmission lines based on DCP-YOLOv8. The method employs deformable convolution (C2f_DCNv3) to enhance the defect feature extraction capability, and designs a re-parameterized cross phase feature fusion structure (RCSP) to optimize and fuse high-level semantic features with low level spatial features, thus improving the capability of the model to recognize defects at different scales while significantly reducing the model parameters; additionally, it combines the dynamic detection head and deformable convolutional v3\'s detection head (DCNv3-Dyhead) to enhance the feature expression capability and the utilization of contextual information to further improve the detection accuracy. Experimental results show that on a dataset containing 20 real transmission line defects, the method increases the average accuracy (mAP@0.5) to 72.2%, an increase of 4.3%, compared with the lightest baseline YOLOv8n model; the number of model parameters is only 2.8 M, a reduction of 9.15%, and the number of processed frames per second (FPS) reaches 103, which meets the real time detection demand. In the scenario of multi type defect detection, it effectively balances detection accuracy and performance with quantitative generalizability.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

公众号