yolo

YOLO
  • 文章类型: Journal Article
    楔形文字,一种古老的艺术风格,让我们看到过去。除了埃及象形文字,楔形文字是最古老的书写系统之一。它出现在公元前四千年的后半期。大多数人认为苏美尔人最初在美索不达米亚南部创造了它。许多历史学家将希伯来语的起源放在古代。例如,我们使用相同的方法来破译楔形文字语言;在学习如何破译一种旧语言之后,我们将拜访一位考古学家,学习如何破译任何其他古代语言。我们提出了一种基于深度学习的标志检测器方法,以加快此过程,以根据希伯来语字母内容识别和分组楔形文字板图像。众所周知,希伯来语字母表很难收集深度学习所需的训练数据,而且成本很高。这需要将希伯来语字符封闭在盒子里。我们通过使用预先存在的音译和以拉丁字符表示平板电脑内容的逐符号表示来解决此问题。我们推荐一种监督的方法,因为这些方法不包括符号定位:我们通过将其与相应的音译进行比较来找到平板电脑照片中的音译符号。然后,使用这些局部化的标志而不是利用注释来重新训练标志检测器。之后,更有效的符号检测器提高了对准质量。因此,本研究旨在使用Yolov8对象识别预训练模型来识别希伯来语字符并对楔形文字片进行分类。说明希伯来语段落的图像已从希伯来语书中剔除。这本书被称为旧约,它被组织成大约500个插图,以帮助阅读和发音的字符。最近在伊拉克发现了另一个古老的文献,追溯到500。经过预处理和增强,它达到了1000多张照片。楔形文字数字图书馆倡议(CDLI)网站和伊拉克博物馆编制了楔形文字板的照片,每种语言都有一千多张照片。
    Cuneiform writing, an old art style, allows us to see into the past. Aside from Egyptian hieroglyphs, the cuneiform script is one of the oldest writing systems. It emerged in the second half of the fourth millennium BC. Most people believe that the Sumerians originally created it in southern Mesopotamia. Many historians place Hebrew\'s origins in antiquity. For example, we used the same approach to decipher the cuneiform languages; after learning how to decipher one old language, we would visit an archaeologist to learn how to decipher any other ancient language. We propose a deep-learning-based sign detector method to speed up this procedure to identify and group cuneiform tablet images according to Hebrew letter content. The Hebrew alphabet is notoriously difficult and costly to gather the training data needed for deep learning, which entails enclosing Hebrew characters in boxes. We solve this problem by using pre-existing transliterations and a sign-by-sign representation of the tablet\'s content in Latin characters. We recommend one of the supervised approaches because these do not include sign localization: We Find the transliteration signs in the tablet photographs by comparing them to their corresponding transliterations. Then, retrain the sign detector using these localized signs instead of utilizing annotations. Afterward, a more effective sign detector enhances the alignment quality. Consequently, this research aims to use the Yolov8 object identification pretraining model to identify Hebrew characters and categorize the cuneiform tablets. Images illustrating Hebrew passages have been culled from a Hebrew-language book. This book is known as the Old Testament, and it was organized into around 500 illustrations to aid in reading and pronouncing the characters. Another ancient document was recently discovered in Iraq, dating back to 500. It reached over a thousand photos after pre-processing and augmentation. The cuneiform digital library initiative (CDLI) website and the Iraqi Museum have compiled photographs of cuneiform tablets, with over a thousand photos available in each language.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    苗圃植物病害的监测,养殖场和果园对维持植物健康至关重要。火疫病(Erwiniaamylovora)仍然是水果生产中最危险的疾病之一,因为它可以传播流行病并造成巨大的经济损失。因此,所有措施都旨在防止病原体在果园中的传播,并在早期阶段包含感染[1-6]。如果可以提高果园疾病监测的空间和时间分辨率,则植物病害控制效率将受益于数字监测系统的开发[7]。在这种情况下,开发了基于RGB图像的果园火灾疫病数字监测系统。在2021年至2024年之间,在不同的天气条件和不同的相机下收集了9个日期的数据。德国的数据源位置是JuliusKühn研究所(JKI)的实验果园,多森海姆水果作物和葡萄栽培植物保护研究所,Quedlinburg的JuliusKühn抗性研究和胁迫耐受性研究所的实验温室和位于德累斯顿-皮尔尼茨的JKI果树育种研究实验园。在人工接种淀粉欧文氏菌后,对不同基因型的苹果进行RGB图像拍摄,包括品种,野生物种和繁殖后代。呈现的ERWIAM数据集包含手动标记的RGB图像,其大小为1280×1280像素的火疫病感染芽,不同发育阶段的花朵和叶子以及没有症状的背景图像。此外,获得了其他植物病害的症状,并将其作为单独的类纳入ERWIAM数据集.每个火灾疫病症状都用计算机视觉注释工具(CVAT[8])使用2点注释(边界框)进行注释,并以YOLO1.1格式(。txt文件)。该数据集总共包含1611个注释图像和87个背景图像。该数据集可以用作研究人员和开发人员的资源,用于植物病害监测的数字系统。
    The monitoring of plant diseases in nurseries, breeding farms and orchards is essential for maintaining plant health. Fire blight (Erwinia amylovora) is still one of the most dangerous diseases in fruit production, as it can spread epidemically and cause enormous economic damage. All measures are therefore aimed at preventing the spread of the pathogen in the orchard and containing an infection at an early stage [1-6]. Efficiency in plant disease control benefits from the development of a digital monitoring system if the spatial and temporal resolution of disease monitoring in orchards can be increased [7]. In this context, a digital disease monitoring system for fire blight based on RGB images was developed for orchards. Between 2021 and 2024, data was collected on nine dates under different weather conditions and with different cameras. The data source locations in Germany were the experimental orchard of the Julius Kühn Institute (JKI), Institute of Plant Protection in Fruit Crops and Viticulture in Dossenheim, the experimental greenhouse of the Julius Kühn Institute for Resistance Research and Stress Tolerance in Quedlinburg and the experimental orchard of the JKI for Breeding Research on Fruit Crops located in Dresden-Pillnitz. The RGB images were taken on different apple genotypes after artificial inoculation with Erwinia amylovora, including cultivars, wild species and progeny from breeding. The presented ERWIAM dataset contains manually labelled RGB images with a size of 1280  × 1280 pixels of fire blight infected shoots, flowers and leaves in different stages of development as well as background images without symptoms. In addition, symptoms of other plant diseases were acquired and integrated into the ERWIAM dataset as a separate class. Each fire blight symptom was annotated with the Computer Vision Annotation Tool (CVAT [8]) using 2-point annotations (bounding boxes) and presented in YOLO 1.1 format (.txt files). The dataset contains a total of 1611 annotated images and 87 background images. This dataset can be used as a resource for researchers and developers working on digital systems for plant disease monitoring.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    近年来,在智能温室领域取得了重大进展,特别是在应用计算机视觉和机器人授粉花。机器人授粉有几个好处,包括减少劳动力需求和通过人工番茄授粉保存昂贵的花粉。然而,以前的研究主要集中在单独标记和检测番茄花。因此,这项研究的目的是开发一种同时标记的综合方法,培训,检测专为机器人授粉而定制的番茄花。为了实现这一点,使用众所周知的模型采用了迁移学习技术,即YOLOv5和最近推出的YOLOv8,用于番茄花的检测。使用相同的图像数据集评估了两个模型的性能,并根据他们的平均精度(AP)得分进行比较,以确定优越的模型。结果表明,YOLOv8在番茄花蕾检测中获得了92.6%的较高平均AP(mAP),表现优于YOLOv5,为91.2%。值得注意的是,YOLOv8还展示了当考虑在检测期间调整大小为640×640像素的1920×1080像素的图像大小时0.7ms的推断速度。在早晨和傍晚时段采集图像数据集,以最小化照明条件对检测模型的影响。这些发现凸显了YOLOv8用于实时检测番茄花和芽的潜力,能够进一步估计花朵盛开的高峰,并促进机器人授粉。在机器人授粉的背景下,该研究还重点介绍了所提出的检测模型在3P2R龙门机器人上的部署。该研究介绍了龙门机器人的运动学模型和改进的电路。在授粉过程中,采用基于位置的视觉伺服方法来接近检测到的花朵。所提出的视觉伺服方法的有效性在实验室环境中的非集群和集群工厂环境中都得到了验证。此外,这项研究为温室系统领域的专家提供了宝贵的理论和实践见解,特别是在使用计算机视觉的花朵检测算法的设计及其在温室中使用的机器人系统中的部署。
    In recent years, significant advancements have been made in the field of smart greenhouses, particularly in the application of computer vision and robotics for pollinating flowers. Robotic pollination offers several benefits, including reduced labor requirements and preservation of costly pollen through artificial tomato pollination. However, previous studies have primarily focused on the labeling and detection of tomato flowers alone. Therefore, the objective of this study was to develop a comprehensive methodology for simultaneously labeling, training, and detecting tomato flowers specifically tailored for robotic pollination. To achieve this, transfer learning techniques were employed using well-known models, namely YOLOv5 and the recently introduced YOLOv8, for tomato flower detection. The performance of both models was evaluated using the same image dataset, and a comparison was made based on their Average Precision (AP) scores to determine the superior model. The results indicated that YOLOv8 achieved a higher mean AP (mAP) of 92.6% in tomato flower and bud detection, outperforming YOLOv5 with 91.2%. Notably, YOLOv8 also demonstrated an inference speed of 0.7 ms when considering an image size of 1920 × 1080 pixels resized to 640 × 640 pixels during detection. The image dataset was acquired during both morning and evening periods to minimize the impact of lighting conditions on the detection model. These findings highlight the potential of YOLOv8 for real-time detection of tomato flowers and buds, enabling further estimation of flower blooming peaks and facilitating robotic pollination. In the context of robotic pollination, the study also focuses on the deployment of the proposed detection model on the 3P2R gantry robot. The study introduces a kinematic model and a modified circuit for the gantry robot. The position-based visual servoing method is employed to approach the detected flower during the pollination process. The effectiveness of the proposed visual servoing approach is validated in both un-clustered and clustered plant environments in the laboratory setting. Additionally, this study provides valuable theoretical and practical insights for specialists in the field of greenhouse systems, particularly in the design of flower detection algorithms using computer vision and its deployment in robotic systems used in greenhouses.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    扁线电机的定子是新能源汽车的核心部件。然而,在涂装过程中实时检测质量缺陷是一个挑战。此外,缺陷的数量很大,单个缺陷的像素很少,这使得难以区分缺陷特征并使准确检测变得更加困难。为了解决这个问题,本文提出了YOLOv8s-DFJA网络。该网络基于YOLOv8,它使用DSFI-HEAD代替原始的检测头,实现任务对齐。它增强了分类任务和定位任务之间的联合特征,提高了网络检测能力。LEFG模块取代了YOLOv8s网络主干中的C2f模块,减少了传统BottleNeck结构带来的冗余参数。它还增强了特征提取和梯度流能力,实现了网络的轻量化。对于这项研究,我们制作了我们自己的关于扁线电机的定子涂层质量数据集。数据增强技术(高斯噪声,调节亮度,等。)丰富了数据集,在某种程度上,提高了YOLOv8s-DFJA的鲁棒性和泛化能力。实验结果表明,与YOLOv8s-DFJA相比,mAP@.5指数上升6.4%,精密度指数上升1.1%,召回指数上升8.1%,FPS指数增加9.8FPS/s,参数降低了3Mb。因此,YOLOv8s-DFJA可以更好地实现对扁线电机定子镀层质量的快速准确检测。
    The stator of a flat wire motor is the core component of new energy vehicles. However, detecting quality defects in the coating process in real-time is a challenge. Moreover, the number of defects is large, and the pixels of a single defect are very few, which make it difficult to distinguish the defect features and make accurate detection more difficult. To solve this problem, this article proposes the YOLOv8s-DFJA network. The network is based on YOLOv8s, which uses DSFI-HEAD to replace the original detection head, realizing task alignment. It enhances joint features between the classification task and localization task and improves the ability of network detection. The LEFG module replaces the C2f module in the backbone of the YOLOv8s network that reduces the redundant parameters brought by the traditional BottleNeck structure. It also enhances the feature extraction and gradient flow ability to achieve the lightweight of the network. For this research, we produced our own dataset of stator coating quality regarding flat wire motors. Data augmentation technology (Gaussian noise, adjusting brightness, etc.) enriches the dataset, to a certain extent, which improves the robustness and generalization ability of YOLOv8s-DFJA. The experimental results show that in the performance of YOLOv8s-DFJA compared with YOLOv8s, the mAP@.5 index increased by 6.4%, the precision index increased by 1.1%, the recall index increased by 8.1%, the FPS index increased by 9.8FPS/s, and the parameters decreased by 3 Mb. Therefore, YOLOv8s-DFJA can be better realize the fast and accurate detection of the stator coating quality of flat wire motors.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    目的:木材包含不同的细胞类型,如纤维,气管和血管,定义其属性。研究细胞的形状,尺寸,和安排在显微镜图像是至关重要的理解木材的特点。通常,这涉及将样品浸渍(浸泡)在溶液中以分离细胞,然后将它们铺在载玻片上,用覆盖广阔区域的显微镜成像,捕获成千上万的细胞。然而,这些细胞经常在图像中聚集和重叠,使用标准的图像处理方法使分割变得困难和耗时。
    结果:在这项工作中,我们开发了一种自动深度学习分割方法,该方法利用一阶段YOLOv8模型对显微镜图像中浸渍的纤维和血管形式的白杨树进行快速准确的分割和表征。该模型可以分析32,640x25,920像素的图像,并展示了有效的细胞检测和分割,实现78%的MAP0.5-0.95。为了评估模型的稳健性,我们检查了已知较长纤维的转基因树系的纤维。结果与以前的手动测量相当。此外,我们创建了一个用于图像分析的用户友好的Web应用程序,并提供了在GoogleColab上使用的代码。
    结论:通过利用YOLOv8的进步,这项工作提供了一种深度学习解决方案,可以对适合实际应用的木材细胞进行有效的量化和分析。
    OBJECTIVE: Wood comprises different cell types, such as fibers, tracheids and vessels, defining its properties. Studying cells\' shape, size, and arrangement in microscopy images is crucial for understanding wood characteristics. Typically, this involves macerating (soaking) samples in a solution to separate cells, then spreading them on slides for imaging with a microscope that covers a wide area, capturing thousands of cells. However, these cells often cluster and overlap in images, making the segmentation difficult and time-consuming using standard image-processing methods.
    RESULTS: In this work, we developed an automatic deep learning segmentation approach that utilizes the one-stage YOLOv8 model for fast and accurate segmentation and characterization of macerated fiber and vessel form aspen trees in microscopy images. The model can analyze 32,640 x 25,920 pixels images and demonstrate effective cell detection and segmentation, achieving a mAP 0.5 - 0.95 of 78 %. To assess the model\'s robustness, we examined fibers from a genetically modified tree line known for longer fibers. The outcomes were comparable to previous manual measurements. Additionally, we created a user-friendly web application for image analysis and provided the code for use on Google Colab.
    CONCLUSIONS: By leveraging YOLOv8\'s advances, this work provides a deep learning solution to enable efficient quantification and analysis of wood cells suitable for practical applications.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    大脑地图集,它提供了基因分布的信息,蛋白质,神经元,或解剖区域,在当代神经科学研究中起着至关重要的作用。为了根据来自不同大脑样本的图像分析这些物质的空间分布,我们经常需要扭曲和注册单个大脑图像到一个标准的大脑模板。然而,翘曲和配准的过程可能导致空间误差,从而严重降低了分析的准确性。为了解决这个问题,我们开发了一种自动方法,用于从FlyCircuit数据库中分割果蝇大脑中的神经质,以获取荧光图像。该技术允许未来的脑图谱研究在个体水平上准确地进行,而不扭曲和对准标准脑模板。我们的方法,LYNSU(通过YOLO定位,通过U-Net分割),包括两个阶段。在第一阶段,我们使用YOLOv7模型快速定位神经痛,并快速提取小规模3D图像作为第二阶段模型的输入。此阶段在Neuropil定位中实现了99.4%的准确率。在第二阶段,我们使用3DU-Net模型来分割神经痛。LYNSU可以使用由来自仅16个大脑的图像组成的小训练集来实现高分割精度。我们展示了LYNSU在六个不同的神经质或结构上,实现与专业手动注释相当的高分割精度与3D交集(IoU)高达0.869。我们的方法只需要大约7s来分割神经纤维,同时实现与人类注释者相似的性能水平。为了演示LYNSU的用例,我们将其应用于FlyCircuit数据库中的所有雌性果蝇大脑,以研究蘑菇体(MB)的不对称性,果蝇的学习中心。我们使用LYNSU分割双侧MB,并比较每个人左右之间的体积。值得注意的是,8703份有效的大脑样本,10.14%的双侧体积差异超过10%。该研究证明了所提出的方法在果蝇大脑的高通量解剖分析和连接组学构建中的潜力。
    The brain atlas, which provides information about the distribution of genes, proteins, neurons, or anatomical regions, plays a crucial role in contemporary neuroscience research. To analyze the spatial distribution of those substances based on images from different brain samples, we often need to warp and register individual brain images to a standard brain template. However, the process of warping and registration may lead to spatial errors, thereby severely reducing the accuracy of the analysis. To address this issue, we develop an automated method for segmenting neuropils in the Drosophila brain for fluorescence images from the FlyCircuit database. This technique allows future brain atlas studies to be conducted accurately at the individual level without warping and aligning to a standard brain template. Our method, LYNSU (Locating by YOLO and Segmenting by U-Net), consists of two stages. In the first stage, we use the YOLOv7 model to quickly locate neuropils and rapidly extract small-scale 3D images as input for the second stage model. This stage achieves a 99.4% accuracy rate in neuropil localization. In the second stage, we employ the 3D U-Net model to segment neuropils. LYNSU can achieve high accuracy in segmentation using a small training set consisting of images from merely 16 brains. We demonstrate LYNSU on six distinct neuropils or structures, achieving a high segmentation accuracy comparable to professional manual annotations with a 3D Intersection-over-Union (IoU) reaching up to 0.869. Our method takes only about 7 s to segment a neuropil while achieving a similar level of performance as the human annotators. To demonstrate a use case of LYNSU, we applied it to all female Drosophila brains from the FlyCircuit database to investigate the asymmetry of the mushroom bodies (MBs), the learning center of fruit flies. We used LYNSU to segment bilateral MBs and compare the volumes between left and right for each individual. Notably, of 8,703 valid brain samples, 10.14% showed bilateral volume differences that exceeded 10%. The study demonstrated the potential of the proposed method in high-throughput anatomical analysis and connectomics construction of the Drosophila brain.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    (1)背景:无人机(UAV)图像中的小物体往往分散在图像的各个区域,比如角落,并且可能被更大的物体阻挡,以及易受图像噪声的影响。此外,由于尺寸小,这些物体在图像中占据有限的区域,导致缺乏有效的检测特征。(2)方法:为了解决无人机影像中的小目标检测问题,我们介绍了一种新的算法,称为高分辨率特征金字塔网络基于Mamba的YOLO(HRMamba-YOLO)。该算法利用了高分辨率网络(HRNet)的优势,EfficientVMamba,和YOLOv8,集成了双空间金字塔池(双SPP)模块,高效曼巴模块(EMM),和融合曼巴模块(FMM),以增强特征提取和捕获上下文信息。此外,一种新的多尺度特征融合网络,高分辨率特征金字塔网络(HRFPN),和FMM改进了特征交互,增强了小目标检测的性能。(3)结果:对于VisDroneDET数据集,与YOLOv8-m相比,该算法的平均平均精度(mAP)提高了4.4%。实验结果表明,HRMamba实现了37.1%的mAP,超过YOLOv8-m3.8%(Dota1.5数据集)。对于UCAS_AOD数据集和DIOR数据集,我们的模型的MAP比YOLOv8-m模型高1.5%和0.3%,分别。公平地说,所有模型均在没有预训练模型的情况下进行训练.(4)结论:本研究不仅突出了HRMamba-YOLO在小目标检测任务中的卓越性能和效率,而且为未来的研究提供了创新的解决方案和有价值的见解。
    (1) Background: Small objects in Unmanned Aerial Vehicle (UAV) images are often scattered throughout various regions of the image, such as the corners, and may be blocked by larger objects, as well as susceptible to image noise. Moreover, due to their small size, these objects occupy a limited area in the image, resulting in a scarcity of effective features for detection. (2) Methods: To address the detection of small objects in UAV imagery, we introduce a novel algorithm called High-Resolution Feature Pyramid Network Mamba-Based YOLO (HRMamba-YOLO). This algorithm leverages the strengths of a High-Resolution Network (HRNet), EfficientVMamba, and YOLOv8, integrating a Double Spatial Pyramid Pooling (Double SPP) module, an Efficient Mamba Module (EMM), and a Fusion Mamba Module (FMM) to enhance feature extraction and capture contextual information. Additionally, a new Multi-Scale Feature Fusion Network, High-Resolution Feature Pyramid Network (HRFPN), and FMM improved feature interactions and enhanced the performance of small object detection. (3) Results: For the VisDroneDET dataset, the proposed algorithm achieved a 4.4% higher Mean Average Precision (mAP) compared to YOLOv8-m. The experimental results showed that HRMamba achieved a mAP of 37.1%, surpassing YOLOv8-m by 3.8% (Dota1.5 dataset). For the UCAS_AOD dataset and the DIOR dataset, our model had a mAP 1.5% and 0.3% higher than the YOLOv8-m model, respectively. To be fair, all the models were trained without a pre-trained model. (4) Conclusions: This study not only highlights the exceptional performance and efficiency of HRMamba-YOLO in small object detection tasks but also provides innovative solutions and valuable insights for future research.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    猕猴桃花的授粉过程对猕猴桃的产量起着至关重要的作用。实现对猕猴桃花的四个阶段的准确快速鉴定对于提高授粉效率至关重要。在这项研究中,提高猕猴桃授粉效率,我们提出了一种新的全阶段猕猴桃花授粉检测算法KIWI-YOLO,基于频域特征的融合。我们的算法利用频域和空间域信息来改善轮廓细节特征的识别,并将决策与上下文信息集成在一起。此外,我们将两级路由注意(BRA)机制与C3相结合,以增强算法对关键领域的关注,导致准确,轻量级,和快速检测。该算法仅用1.8M参数就实现了91.6%的mAP0.5,女班和男班的AP分别达到95%和93.5%,改善了3.8%,1.2%,与原算法相比为6.2%。此外,算法的召回和F1分数分别提高了5.5%和3.1%,分别。此外,我们的模型在检测速度方面表现出显著的优势,只需0.016s处理图像。实验结果表明,本文提出的算法模型能够较好地辅助精准农业生产过程中猕猴桃的授粉,有助于猕猴桃产业的发展。
    The pollination process of kiwifruit flowers plays a crucial role in kiwifruit yield. Achieving accurate and rapid identification of the four stages of kiwifruit flowers is essential for enhancing pollination efficiency. In this study, to improve the efficiency of kiwifruit pollination, we propose a novel full-stage kiwifruit flower pollination detection algorithm named KIWI-YOLO, based on the fusion of frequency-domain features. Our algorithm leverages frequency-domain and spatial-domain information to improve recognition of contour-detailed features and integrates decision-making with contextual information. Additionally, we incorporate the Bi-Level Routing Attention (BRA) mechanism with C3 to enhance the algorithm\'s focus on critical areas, resulting in accurate, lightweight, and fast detection. The algorithm achieves a m A P 0.5 of 91.6% with only 1.8M parameters, the AP of the Female class and the Male class reaches 95% and 93.5%, which is an improvement of 3.8%, 1.2%, and 6.2% compared with the original algorithm. Furthermore, the Recall and F1-score of the algorithm are enhanced by 5.5% and 3.1%, respectively. Moreover, our model demonstrates significant advantages in detection speed, taking only 0.016s to process an image. The experimental results show that the algorithmic model proposed in this study can better assist the pollination of kiwifruit in the process of precision agriculture production and help the development of the kiwifruit industry.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    对钢筋混凝土(RC)结构中的钢筋进行准确检查至关重要,需要仔细计数。利用物体检测的深度学习算法可以通过无人机(UAV)图像促进这一过程。然而,它们的有效性取决于大的可用性,多样化,和标记良好的数据集。本文详细介绍了使用基于深度学习的对象检测方法创建专门用于计数钢筋的数据集。数据集包括874张原始图像,分为三个子集:524张用于训练的图像(60%),175用于验证(20%),和175用于测试(20%)。为了增强训练数据,我们应用了八种增强技术——亮度,对比,透视,旋转,scale,剪切,翻译,并且只模糊到训练子集。这导致了九个不同的数据集:每个增强技术一个,一个组合了增强集中的所有技术。专家注释器以VOCXML格式标记数据集。虽然这项研究的重点是钢筋计数,原始数据集可以适应其他任务,如估计钢筋直径或分类钢筋形状,通过提供必要的注释。
    Accurate inspection of rebars in Reinforced Concrete (RC) structures is essential and requires careful counting. Deep learning algorithms utilizing object detection can facilitate this process through Unmanned Aerial Vehicle (UAV) imagery. However, their effectiveness depends on the availability of large, diverse, and well-labelled datasets. This article details the creation of a dataset specifically for counting rebars using deep learning-based object detection methods. The dataset comprises 874 raw images, divided into three subsets: 524 images for training (60 %), 175 for validation (20 %), and 175 for testing (20 %). To enhance the training data, we applied eight augmentation techniques-brightness, contrast, perspective, rotation, scale, shearing, translation, and blurring-exclusively to the training subset. This resulted in nine distinct datasets: one for each augmentation technique and one combining all techniques in augmentation sets. Expert annotators labelled the dataset in VOC XML format. While this research focuses on rebar counting, the raw dataset can be adapted for other tasks, such as estimating rebar diameter or classifying rebar shapes, by providing the necessary annotations.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    视觉目标检测技术的广泛研究和实现,极大地改变了自动驾驶行业。自动驾驶在很大程度上依赖于视觉传感器来感知和分析环境。然而,在极端天气条件下,如大雨,雾,或者低光,这些传感器可能会遇到干扰,导致图像质量下降和检测精度下降,从而增加了自动驾驶的风险。为了应对这些挑战,我们提出了自适应图像增强(AIE)-YOLO,一种新的目标检测方法,以提高极端天气条件下的道路目标检测精度。为了解决极端天气下图像质量下降的问题,设计了一种改进的自适应图像增强模块。该模块根据不同的场景条件动态调整道路图像的像素特征,从而增强物体的可见度并抑制不相关的背景干扰。此外,我们引入了空间特征提取模块,以在复杂背景下自适应地增强模型的空间建模能力。此外,设计了通道特征提取模块,以自适应地增强模型的表示和泛化能力。由于难以获得各种极端天气条件的真实世界数据,我们构建了一个新的基准数据集,称为极端天气模拟-稀有对象数据集。该数据集包括十种类型的模拟极端天气场景,并建立在公开可用的稀有物体检测数据集上。在极端天气模拟稀有对象数据集上进行的大量实验表明,AIE-YOLO优于现有的最新方法,在极端天气条件下实现出色的检测性能。
    The widespread research and implementation of visual object detection technology have significantly transformed the autonomous driving industry. Autonomous driving relies heavily on visual sensors to perceive and analyze the environment. However, under extreme weather conditions, such as heavy rain, fog, or low light, these sensors may encounter disruptions, resulting in decreased image quality and reduced detection accuracy, thereby increasing the risk for autonomous driving. To address these challenges, we propose adaptive image enhancement (AIE)-YOLO, a novel object detection method to enhance road object detection accuracy under extreme weather conditions. To tackle the issue of image quality degradation in extreme weather, we designed an improved adaptive image enhancement module. This module dynamically adjusts the pixel features of road images based on different scene conditions, thereby enhancing object visibility and suppressing irrelevant background interference. Additionally, we introduce a spatial feature extraction module to adaptively enhance the model\'s spatial modeling capability under complex backgrounds. Furthermore, a channel feature extraction module is designed to adaptively enhance the model\'s representation and generalization abilities. Due to the difficulty in acquiring real-world data for various extreme weather conditions, we constructed a novel benchmark dataset named extreme weather simulation-rare object dataset. This dataset comprises ten types of simulated extreme weather scenarios and is built upon a publicly available rare object detection dataset. Extensive experiments conducted on the extreme weather simulation-rare object dataset demonstrate that AIE-YOLO outperforms existing state-of-the-art methods, achieving excellent detection performance under extreme weather conditions.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

公众号