Object detection

对象检测
  • 文章类型: Journal Article
    背景:成熟度是一种显著影响水果品质的表型,构成种植和收获过程中的关键因素。手动检测方法和实验分析,然而,效率低下,成本高昂。
    结果:在这项研究中,提出了一种轻量高效的甜瓜成熟度检测方法,MRD-YOLO,基于改进的目标检测算法。该方法结合了轻量级骨干网,MobileNetV3,一种设计范式,协调注意力机制。此外,我们创建了一个基于成熟度的温室的大规模甜瓜数据集。此数据集包含在现场环境中遇到的常见复杂性,如闭塞,重叠,和不同的光强度。MRD-YOLO在该数据集上实现了97.4%的平均精度,实现准确可靠的甜瓜成熟度检测。此外,该方法只需要4.8GFLOP和2.06M参数,代表基线YOLOv8n模型的58.5%和68.4%,分别。它在平衡精度和计算效率方面全面优于现有方法。此外,它在GPU环境中保持实时推理能力,并在CPU环境中演示出众的推理速度。MRD-YOLO的轻量级设计预计将部署在各种资源受限的移动和边缘设备中,比如挑选机器人。特别值得注意的是,在从Roboflow平台获得的两个甜瓜数据集上测试时,它的性能,平均精度达到85.9%。这强调了其对未经训练的数据的出色泛化能力。
    结论:本研究提出了一种检测甜瓜成熟度的有效方法,以及这项研究中使用的数据集,除了检测方法之外,将为各类水果的成熟度检测提供有价值的参考。
    BACKGROUND: Ripeness is a phenotype that significantly impacts the quality of fruits, constituting a crucial factor in the cultivation and harvesting processes. Manual detection methods and experimental analysis, however, are inefficient and costly.
    RESULTS: In this study, we propose a lightweight and efficient melon ripeness detection method, MRD-YOLO, based on an improved object detection algorithm. The method combines a lightweight backbone network, MobileNetV3, a design paradigm Slim-neck, and a Coordinate Attention mechanism. Additionally, we have created a large-scale melon dataset sourced from a greenhouse based on ripeness. This dataset contains common complexities encountered in the field environment, such as occlusions, overlapping, and varying light intensities. MRD-YOLO achieves a mean Average Precision of 97.4% on this dataset, achieving accurate and reliable melon ripeness detection. Moreover, the method demands only 4.8 G FLOPs and 2.06 M parameters, representing 58.5% and 68.4% of the baseline YOLOv8n model, respectively. It comprehensively outperforms existing methods in terms of balanced accuracy and computational efficiency. Furthermore, it maintains real-time inference capability in GPU environments and demonstrates exceptional inference speed in CPU environments. The lightweight design of MRD-YOLO is anticipated to be deployed in various resource constrained mobile and edge devices, such as picking robots. Particularly noteworthy is its performance when tested on two melon datasets obtained from the Roboflow platform, achieving a mean Average Precision of 85.9%. This underscores its excellent generalization ability on untrained data.
    CONCLUSIONS: This study presents an efficient method for melon ripeness detection, and the dataset utilized in this study, alongside the detection method, will provide a valuable reference for ripeness detection across various types of fruits.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    遥感图像中感兴趣目标的精确定位对目标识别具有重要意义,资源管理,决策和救灾响应。然而,许多困难,像复杂的背景,密集的目标数量,大规模的变化,和小规模的物体,这使得检测精度不能令人满意。为了提高检测精度,我们提出了一种自适应相邻上下文协商网络(A2CN-Net)。首先,提出了复合快速傅里叶卷积(CFFC)模块,以减少小物体的信息损失,将其插入骨干网络以获取频谱全球上下文信息。然后,全局上下文信息增强(GCIE)模块用于捕获和聚合全局空间特征,这有利于定位不同尺度的对象。此外,为了减轻相邻特征层融合引起的混叠效应,给出了一种新的自适应相邻上下文协商网络(A2CN),用于多级特征的自适应集成,由本地和相邻分支组成,局部分支自适应突出特征信息,相邻分支在相邻级别引入全局信息以增强特征表示。同时,考虑到不同维度特征层焦点的可变性,将可学习的权重应用于局部和相邻分支以进行自适应特征融合。最后,在几个可用的公共数据集中进行了广泛的实验,包括DIOR和DOTA-V1.0。实验研究表明,A2CN-Net可以显著提高检测性能,随着MAP增加到74.2%和79.2%,分别。
    Accurate localization of objects of interest in remote sensing images (RSIs) is of great significance for object identification, resource management, decision-making and disaster relief response. However, many difficulties, like complex backgrounds, dense target quantities, large-scale variations, and small-scale objects, which make the detection accuracy unsatisfactory. To improve the detection accuracy, we propose an Adaptive Adjacent Context Negotiation Network (A2CN-Net). Firstly, the composite fast Fourier convolution (CFFC) module is given to reduce the information loss of small objects, which is inserted into the backbone network to obtain spectral global context information. Then, the Global Context Information Enhancement (GCIE) module is given to capture and aggregate global spatial features, which is beneficial for locating objects of different scales. Furthermore, to alleviate the aliasing effect caused by the fusion of adjacent feature layers, a novel Adaptive Adjacent Context Negotiation network (A2CN) is given to adaptive integration of multi-level features, which consists of local and adjacent branches, with the local branch adaptively highlighting feature information and the adjacent branch introducing global information at the adjacent level to enhance feature representation. In the meantime, considering the variability in the focus of feature layers in different dimensions, learnable weights are applied to the local and adjacent branches for adaptive feature fusion. Finally, extensive experiments are performed in several available public datasets, including DIOR and DOTA-v1.0. Experimental studies show that A2CN-Net can significantly boost detection performance, with mAP increasing to 74.2% and 79.2%, respectively.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    细菌图像分析在各个领域中起着至关重要的作用,为研究细菌结构生物学提供有价值的信息和见解,诊断和治疗由病原菌引起的传染病,发现和开发可以对抗细菌感染的药物,等。因此,它促使人们努力自动化细菌图像分析任务。通过自动化分析任务和利用更先进的计算技术,例如深度学习(DL)算法,细菌图像分析有助于快速,更准确,高效,可靠,和标准化分析,导致增进理解,诊断,和控制细菌相关的现象。
    DL算法的三个目标检测网络,即SSD-MobileNetV2,EfficientDet,和YOLOv4,被开发用于自动检测大肠杆菌(E.大肠杆菌)显微镜图像中的细菌。开发了多任务DL框架,以根据其各自的生长阶段对细菌进行分类,其中包括杆状细胞,分裂的细胞,和微菌落。在训练目标检测模型之前进行数据预处理步骤,包括图像增强,图像注释,和数据拆分。使用基于平均精度(MAP)的定量评估方法评估DL技术的性能,精度,召回,和F1得分。对模型的性能指标进行了比较和分析。然后选择最佳的DL模型来执行识别杆状细胞的多任务对象检测,分裂的细胞,和微菌落。
    从三个建议的DL模型生成的测试图像的输出显示出较高的检测精度,YOLOv4实现检测的最高置信度得分范围,并能够为大肠杆菌细菌的不同生长阶段创建不同颜色的边界框。在统计分析方面,在提出的三个模型中,YOLOv4表现出卓越的性能,以最高的精度实现98%的最高MAP,召回,F1得分为86%,97%,91%,分别。
    这项研究证明了其有效性,潜力,以及DL方法在多任务细菌图像分析中的适用性,专注于从显微图像中自动化细菌的检测和分类。所提出的模型可以输出带有包围每个检测到的大肠杆菌细菌的边界框的图像,标记为它们的生长阶段和检测的置信水平。所有提出的目标检测模型都取得了有希望的结果,YOLOv4优于其他型号。
    UNASSIGNED: Bacterial image analysis plays a vital role in various fields, providing valuable information and insights for studying bacterial structural biology, diagnosing and treating infectious diseases caused by pathogenic bacteria, discovering and developing drugs that can combat bacterial infections, etc. As a result, it has prompted efforts to automate bacterial image analysis tasks. By automating analysis tasks and leveraging more advanced computational techniques, such as deep learning (DL) algorithms, bacterial image analysis can contribute to rapid, more accurate, efficient, reliable, and standardised analysis, leading to enhanced understanding, diagnosis, and control of bacterial-related phenomena.
    UNASSIGNED: Three object detection networks of DL algorithms, namely SSD-MobileNetV2, EfficientDet, and YOLOv4, were developed to automatically detect Escherichia coli (E. coli) bacteria from microscopic images. The multi-task DL framework is developed to classify the bacteria according to their respective growth stages, which include rod-shaped cells, dividing cells, and microcolonies. Data preprocessing steps were carried out before training the object detection models, including image augmentation, image annotation, and data splitting. The performance of the DL techniques is evaluated using the quantitative assessment method based on mean average precision (mAP), precision, recall, and F1-score. The performance metrics of the models were compared and analysed. The best DL model was then selected to perform multi-task object detections in identifying rod-shaped cells, dividing cells, and microcolonies.
    UNASSIGNED: The output of the test images generated from the three proposed DL models displayed high detection accuracy, with YOLOv4 achieving the highest confidence score range of detection and being able to create different coloured bounding boxes for different growth stages of E. coli bacteria. In terms of statistical analysis, among the three proposed models, YOLOv4 demonstrates superior performance, achieving the highest mAP of 98% with the highest precision, recall, and F1-score of 86%, 97%, and 91%, respectively.
    UNASSIGNED: This study has demonstrated the effectiveness, potential, and applicability of DL approaches in multi-task bacterial image analysis, focusing on automating the detection and classification of bacteria from microscopic images. The proposed models can output images with bounding boxes surrounding each detected E. coli bacteria, labelled with their growth stage and confidence level of detection. All proposed object detection models have achieved promising results, with YOLOv4 outperforming the other models.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    背景:术后甲状旁腺功能减退是甲状腺切除术的主要并发症,当甲状旁腺在手术过程中意外受损时发生。尽管术中图像由于其复杂的性质而很少用于训练人工智能(AI),可以训练AI使用各种增强方法在术中检测甲状旁腺。这项研究的目的是训练一种有效的AI模型来检测甲状腺切除术期间的甲状旁腺。
    方法:在甲状腺叶切除术过程中收集甲状旁腺的视频剪辑。确认的甲状旁腺图像用于根据增强状态训练三种类型的数据集:基线,几何变换,以及基于生成对抗网络的图像修复。主要结果是AI检测甲状旁腺的平均精确度。
    结果:152从150例接受单侧肺叶切除术的患者中获得了细针抽吸证实的甲状旁腺图像。基于基线数据的AI模型检测甲状旁腺的平均精确度为77%。通过应用几何变换和图像修补增强方法来增强这种性能,几何变换数据增强数据集显示出比图像修复模型(78.6%)更高的平均精度(79%)。当使用完全不同的甲状腺切除术方法对该模型进行外部验证时,图像修补方法(46%)比几何变换(37%)和基线(33%)方法更有效.
    结论:发现该AI模型是甲状腺切除术中甲状旁腺识别的有效且可推广的工具,特别是在适当的增强方法的帮助下。比较模型性能和外科医生识别的其他研究,然而,需要评估这个人工智能模型的真正临床相关性。
    BACKGROUND: Postoperative hypoparathyroidism is a major complication of thyroidectomy, occurring when the parathyroid glands are inadvertently damaged during surgery. Although intraoperative images are rarely used to train artificial intelligence (AI) because of its complex nature, AI may be trained to intraoperatively detect parathyroid glands using various augmentation methods. The purpose of this study was to train an effective AI model to detect parathyroid glands during thyroidectomy.
    METHODS: Video clips of the parathyroid gland were collected during thyroid lobectomy procedures. Confirmed parathyroid images were used to train three types of datasets according to augmentation status: baseline, geometric transformation, and generative adversarial network-based image inpainting. The primary outcome was the average precision of the performance of AI in detecting parathyroid glands.
    RESULTS: 152 Fine-needle aspiration-confirmed parathyroid gland images were acquired from 150 patients who underwent unilateral lobectomy. The average precision of the AI model in detecting parathyroid glands based on baseline data was 77%. This performance was enhanced by applying both geometric transformation and image inpainting augmentation methods, with the geometric transformation data augmentation dataset showing a higher average precision (79%) than the image inpainting model (78.6%). When this model was subjected to external validation using a completely different thyroidectomy approach, the image inpainting method was more effective (46%) than both the geometric transformation (37%) and baseline (33%) methods.
    CONCLUSIONS: This AI model was found to be an effective and generalizable tool in the intraoperative identification of parathyroid glands during thyroidectomy, especially when aided by appropriate augmentation methods. Additional studies comparing model performance and surgeon identification, however, are needed to assess the true clinical relevance of this AI model.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    药物泡罩包装中的缺陷检测是在制造时检测片剂中出现的缺陷时获得准确结果的最具挑战性的任务。传统的缺陷检测方法包括人为干预以检查泡罩包装内的片剂质量,这是低效的,耗时,增加劳动力成本。为了缓解这个问题,YOLO系列主要用于许多行业,用于连续生产中的实时缺陷检测。为了增强特征提取能力并减少实时环境中的计算开销,CBS-YOLOv8是通过增强YOLOv8模型提出的。在所提出的CBS-YOLOv8中,引入了协调关注以通过捕获空间和跨信道信息并且还保持远程依赖性来提高特征提取能力。在YOLOv8中还引入了BiFPN(加权双向特征金字塔网络),以增强每个卷积层的特征融合,以避免更精确的信息丢失。通过实施SimSPPF(简单空间金字塔快速池化),提高了模型的效率,这降低了计算需求和模型复杂性,从而提高了速度。包含有缺陷的平板电脑图像的自定义数据集用于训练所提出的模型。然后通过将CBS-YOLOv8模型与各种其他模型进行比较来评估其性能。在自定义数据集上的实验结果表明,CBS-YOLOv8模型实现了97.4%的mAP和79.25FPS的推理速度,表现优于其他型号。所提出的模型还在SESOVERA-ST盐水瓶填充水平监测数据集上进行了评估,达到了99.3%的mAP50。这表明CBS-YOLOv8提供了一个优化的检测过程,能够及时发现和纠正缺陷,从而加强制造环境中的质量保证实践。
    Defect detection in pharmaceutical blister packages is the most challenging task to get an accurate result in detecting defects that arise in tablets while manufacturing. Conventional defect detection methods include human intervention to check the quality of tablets within the blister packages, which is inefficient, time-consuming, and increases labor costs. To mitigate this issue, the YOLO family is primarily used in many industries for real-time defect detection in continuous production. To enhance the feature extraction capability and reduce the computational overhead in a real-time environment, the CBS-YOLOv8 is proposed by enhancing the YOLOv8 model. In the proposed CBS-YOLOv8, coordinate attention is introduced to improve the feature extraction capability by capturing the spatial and cross-channel information and also maintaining the long-range dependencies. The BiFPN (weighted bi-directional feature pyramid network) is also introduced in YOLOv8 to enhance the feature fusion at each convolution layer to avoid more precise information loss. The model\'s efficiency is enhanced through the implementation of SimSPPF (simple spatial pyramid pooling fast), which reduces computational demands and model complexity, resulting in improved speed. A custom dataset containing defective tablet images is used to train the proposed model. The performance of the CBS-YOLOv8 model is then evaluated by comparing it with various other models. Experimental results on the custom dataset reveal that the CBS-YOLOv8 model achieves a mAP of 97.4% and an inference speed of 79.25 FPS, outperforming other models. The proposed model is also evaluated on SESOVERA-ST saline bottle fill level monitoring dataset achieved the mAP50 of 99.3%. This demonstrates that CBS-YOLOv8 provides an optimized inspection process, enabling prompt detection and correction of defects, thus bolstering quality assurance practices in manufacturing settings.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    为了研究植物器官,有必要研究植物的三维(3D)结构。近年来,通过计算机断层扫描(CT)进行的无损测量已用于了解植物的3D结构。在这项研究中,我们以菊花小头花序为例,重点研究了3D小头花序芽结构中容器和小花之间的接触点,以研究小花在容器上的3D排列。要确定接触点的3D顺序,我们从CT体积数据构建了切片图像,并检测了图像中的容器和小花。然而,因为每个CT样本都包含数百个待处理的切片图像,每个C.seticuspe头花序都包含几个小花,手动检测容器和小花是劳动密集型的。因此,利用图像识别技术,提出了一种基于CT切片图像的接触点自动检测方法。所提出的方法使用接触点仅存在于插座周围的先验知识来提高接触点检测的准确性。此外,检测结果的积分使得能够估计接触点的3D位置。根据实验结果,我们证实了所提出的方法可以高精度地检测切片图像上的接触,并通过聚类估计它们的3D位置。此外,与样本无关的实验表明,所提出的方法达到了与样本相关实验相同的检测精度。
    To study plant organs, it is necessary to investigate the three-dimensional (3D) structures of plants. In recent years, non-destructive measurements through computed tomography (CT) have been used to understand the 3D structures of plants. In this study, we use the Chrysanthemum seticuspe capitulum inflorescence as an example and focus on contact points between the receptacles and florets within the 3D capitulum inflorescence bud structure to investigate the 3D arrangement of the florets on the receptacle. To determine the 3D order of the contact points, we constructed slice images from the CT volume data and detected the receptacles and florets in the image. However, because each CT sample comprises hundreds of slice images to be processed and each C. seticuspe capitulum inflorescence comprises several florets, manually detecting the receptacles and florets is labor-intensive. Therefore, we propose an automatic contact point detection method based on CT slice images using image recognition techniques. The proposed method improves the accuracy of contact point detection using prior knowledge that contact points exist only around the receptacle. In addition, the integration of the detection results enables the estimation of the 3D position of the contact points. According to the experimental results, we confirmed that the proposed method can detect contacts on slice images with high accuracy and estimate their 3D positions through clustering. Additionally, the sample-independent experiments showed that the proposed method achieved the same detection accuracy as sample-dependent experiments.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    牛奶蓟,水飞蓟草(L.),由于水飞蓟素含量高,是一种众所周知的用于治疗肝病的药用植物。种子含有elaiosome,附着在种子上的肉质结构,它被认为是包括水飞蓟素在内的许多代谢物的丰富来源。仅使用图像分析很难分割elaiosomes,这使得无法量化elaiosome表型。本研究提出了一种使用Detectron2深度学习算法对水飞蓟籽中的eliosomes进行半自动检测和分割的新方法。使用一百张手动标记的图像来训练初始的elaiosome检测模型。该模型用于从新数据集预测elaiosome,精确的预测是手动选择的,并用作新的标记图像,用于重新训练模型。这样的半自动图像标记,即,利用前一阶段的预测结果对模型进行再训练,允许产生足够的标记数据进行重新训练。最后,总共使用6,000个标记图像来训练Detectron2以进行elaiosome检测,并获得了有希望的结果。结果表明,Detectron2在检测水飞蓟籽elaiosome方面的有效性,准确率为99.9%。所提出的方法可以自动检测和分割水飞蓟籽中的elaiosome。通过ImageJ中基于图像的高通量表型分析,使用预测的elaiosome的掩模图像来分析其作为种子表型性状之一的面积以及其他种子形态性状。实现elaiosome和其他种子形态性状的高通量表型将有助于育种具有理想性状的水飞蓟品种。
    Milk thistle, Silybum marianum (L.), is a well-known medicinal plant used for the treatment of liver diseases due to its high content of silymarin. The seeds contain elaiosome, a fleshy structure attached to the seeds, which is believed to be a rich source of many metabolites including silymarin. Segmentation of elaiosomes using only image analysis is difficult, and this makes it impossible to quantify the elaiosome phenotypes. This study proposes a new approach for semi-automated detection and segmentation of elaiosomes in milk thistle seed using the Detectron2 deep learning algorithm. One hundred manually labeled images were used to train the initial elaiosome detection model. This model was used to predict elaiosome from new datasets, and the precise predictions were manually selected and used as new labeled images for retraining the model. Such semi-automatic image labeling, i.e., using the prediction results of the previous stage for retraining the model, allowed the production of sufficient labeled data for retraining. Finally, a total of 6,000 labeled images were used to train Detectron2 for elaiosome detection and attained a promising result. The results demonstrate the effectiveness of Detectron2 in detecting milk thistle seed elaiosomes with an accuracy of 99.9%. The proposed method automatically detects and segments elaiosome from the milk thistle seed. The predicted mask images of elaiosome were used to analyze its area as one of the seed phenotypic traits along with other seed morphological traits by image-based high-throughput phenotyping in ImageJ. Enabling high-throughput phenotyping of elaiosome and other seed morphological traits will be useful for breeding milk thistle cultivars with desirable traits.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    牙周炎的严重程度可以通过计算牙槽骨和牙骨质-牙釉质交界处(CEJ)之间的牙槽峰(ALC)水平和骨丢失水平来分析。然而,牙医需要在根尖周X光片(PA)上手动标记症状以评估骨质流失,一个既耗时又容易出错的过程。这项研究提出了以下新方法,有助于疾病的评估并减少错误。首先,采用创新的牙周炎图像增强方法来提高PA图像质量。随后,目标检测可以从PA图像中准确提取单颗牙齿,最高准确率为97.01%。在这项研究中开发的实例分割准确地提取了感兴趣的区域,能够生成牙骨和牙冠面罩,准确率分别为93.48%和96.95%。最后,提出了一种新的检测算法来自动标记有症状牙齿的CEJ和ALC,促进牙医更快地准确评估骨质流失的严重程度。本研究中使用的PA图像数据库,长贡医疗中心提供的IRB编号为02002030B0,台湾,通过这项研究中开发的技术,显着减少了牙科诊断所需的时间,并提高了医疗保健质量。
    The severity of periodontitis can be analyzed by calculating the loss of alveolar crest (ALC) level and the level of bone loss between the tooth\'s bone and the cemento-enamel junction (CEJ). However, dentists need to manually mark symptoms on periapical radiographs (PAs) to assess bone loss, a process that is both time-consuming and prone to errors. This study proposes the following new method that contributes to the evaluation of disease and reduces errors. Firstly, innovative periodontitis image enhancement methods are employed to improve PA image quality. Subsequently, single teeth can be accurately extracted from PA images by object detection with a maximum accuracy of 97.01%. An instance segmentation developed in this study accurately extracts regions of interest, enabling the generation of masks for tooth bone and tooth crown with accuracies of 93.48% and 96.95%. Finally, a novel detection algorithm is proposed to automatically mark the CEJ and ALC of symptomatic teeth, facilitating faster accurate assessment of bone loss severity by dentists. The PA image database used in this study, with the IRB number 02002030B0 provided by Chang Gung Medical Center, Taiwan, significantly reduces the time required for dental diagnosis and enhances healthcare quality through the techniques developed in this research.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    柠檬,作为一种具有丰富营养价值的重要经济作物,在全球范围内具有重要的种植重要性和市场需求。然而,柠檬病害严重影响柠檬的品质和产量,必须及早发现以进行有效控制。本文通过收集柠檬疾病的数据集来满足这一需求,由在不同光照水平下拍摄的726张图像组成,生长阶段,射击距离和疾病状况。通过裁剪高分辨率图像,数据集扩展到2022年的图像,包括4441只健康柠檬和718只患病柠檬,每个图像大约有1-6个目标。然后,我们提出了一种新的模型柠檬表面病YOLO(LSD-YOLO),集成了可切换Atrous卷积(SAConv)和卷积块注意力模块(CBAM),同时设计了C2f-SAC,增加了小目标探测层,增强了关键特征的提取和不同尺度特征的融合。实验结果表明,所提出的LSD-YOLO在收集的数据集上达到了90.62%的精度,MAP@50-95达到80.84%。与原来的YOLOv8n型号相比,mAP@50和mAP@50-95指标都得到了增强。因此,本研究中提出的LSD-YOLO模型提供了对健康和患病柠檬的更准确识别,有助于有效解决柠檬病检测问题。
    Lemon, as an important cash crop with rich nutritional value, holds significant cultivation importance and market demand worldwide. However, lemon diseases seriously impact the quality and yield of lemons, necessitating their early detection for effective control. This paper addresses this need by collecting a dataset of lemon diseases, consisting of 726 images captured under varying light levels, growth stages, shooting distances and disease conditions. Through cropping high-resolution images, the dataset is expanded to 2022 images, comprising 4441 healthy lemons and 718 diseased lemons, with approximately 1-6 targets per image. Then, we propose a novel model lemon surface disease YOLO (LSD-YOLO), which integrates Switchable Atrous Convolution (SAConv) and Convolutional Block Attention Module (CBAM), along with the design of C2f-SAC and the addition of a small-target detection layer to enhance the extraction of key features and the fusion of features at different scales. The experimental results demonstrate that the proposed LSD-YOLO achieves an accuracy of 90.62% on the collected datasets, with mAP@50-95 reaching 80.84%. Compared with the original YOLOv8n model, both mAP@50 and mAP@50-95 metrics are enhanced. Therefore, the LSD-YOLO model proposed in this study provides a more accurate recognition of healthy and diseased lemons, contributing effectively to solving the lemon disease detection problem.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    行人轨迹预测对于开发自动驾驶系统中的防撞算法至关重要,旨在根据检测到的行人过去的轨迹预测他们未来的运动。传统的行人轨迹预测方法涉及一系列的任务,包括检测和跟踪,以收集观察到的行人的历史运动。因此,轨迹预测的准确性在很大程度上依赖于检测和跟踪模型的准确性,让它容易受到他们的表现。先前在轨迹预测方面的研究主要使用公共数据集评估模型性能,这往往忽略了源于检测和跟踪模型的错误。这种监督未能捕捉到不可避免的检测和跟踪不准确的现实场景。在这项研究中,我们研究了集成检测中错误的累积效应,跟踪,和轨迹预测管道。通过实证分析,我们检查了在管道的每个阶段引入的误差,并评估了它们对轨迹预测准确性的集体影响。我们在台湾收集的各种自定义数据集上评估这些模型,以提供全面的评估。我们对这些集成管道得出的结果的分析阐明了检测和跟踪误差对下游任务的重大影响,如轨迹预测和距离估计。
    Pedestrian trajectory prediction is crucial for developing collision avoidance algorithms in autonomous driving systems, aiming to predict the future movement of the detected pedestrians based on their past trajectories. The traditional methods for pedestrian trajectory prediction involve a sequence of tasks, including detection and tracking to gather the historical movement of the observed pedestrians. Consequently, the accuracy of trajectory prediction heavily relies on the accuracy of the detection and tracking models, making it susceptible to their performance. The prior research in trajectory prediction has mainly assessed the model performance using public datasets, which often overlook the errors originating from detection and tracking models. This oversight fails to capture the real-world scenario of inevitable detection and tracking inaccuracies. In this study, we investigate the cumulative effect of errors within integrated detection, tracking, and trajectory prediction pipelines. Through empirical analysis, we examine the errors introduced at each stage of the pipeline and assess their collective impact on the trajectory prediction accuracy. We evaluate these models across various custom datasets collected in Taiwan to provide a comprehensive assessment. Our analysis of the results derived from these integrated pipelines illuminates the significant influence of detection and tracking errors on downstream tasks, such as trajectory prediction and distance estimation.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

公众号