machine vision

机器视觉
  • 文章类型: Journal Article
    人参的地理来源显著影响其营养价值和化学成分,进而影响其市场价格。分析这些差异的传统方法通常很耗时,并且需要大量的试剂,使他们效率低下。因此,高光谱成像(HSI)结合X射线技术用于人参产地的快速和无损可追溯性。最初,通过采用组合的孤立森林算法和密度峰值聚类(DPC)算法,可以有效地拒绝离群样本。随后,利用高光谱数据构建随机森林(RF)和支持向量机(SVM)分类模型。通过应用72种预处理方法及其组合,进一步优化了这些模型。此外,为了提高模型的性能,采用了四种变量筛选算法:SelectKBest,遗传算法(GA),最小绝对收缩和选择运算符(LASSO),和排列特征重要性(PFI)。优化后的模型,利用二阶导数,自动缩放,排列特征重要性,和支持向量机(2ndDer-AS-PFI-SVM),实现了93.4%的预测精度,在独立预测集上,Kappa值为0.876,Brier评分为0.030,F1评分为0.932,AUC为0.994。此外,从彩色和X射线图像中提取的图像数据(包括颜色信息和纹理信息)用于构建分类模型并评估其性能。其中,利用X射线图像的纹理信息构建的SVM模型表现最好,在验证集上实现了63.0%的预测精度,Brier评分为0.181,F1评分为0.518,AUC为0.553。通过基于Stacking策略实现中层融合和高层数据融合,发现采用高光谱光谱信息和X射线图像纹理信息的高级融合的模型明显优于仅使用高光谱光谱信息的模型。这种先进的模型达到了95.2%的预测精度,在独立预测集上,Kappa值为0.912,Brier评分为0.027,F1评分为0.952,AUC为0.997。总之,本研究不仅为人参产地的快速、无损溯源提供了一条新的技术路径,同时也展示了HSI和X射线技术在医药和食品可追溯性领域的联合应用的巨大潜力。
    The geographical origin of Panax ginseng significantly influences its nutritional value and chemical composition, which in turn affects its market price. Traditional methods for analyzing these differences are often time-consuming and require substantial quantities of reagents, rendering them inefficient. Therefore, hyperspectral imaging (HSI) in conjunction with X-ray technology were used for the swift and non-destructive traceability of Panax ginseng origin. Initially, outlier samples were effectively rejected by employing a combined isolated forest algorithm and density peak clustering (DPC) algorithm. Subsequently, random forest (RF) and support vector machine (SVM) classification models were constructed using hyperspectral spectral data. These models were further optimized through the application of 72 preprocessing methods and their combinations. Additionally, to enhance the model\'s performance, four variable screening algorithms were employed: SelectKBest, genetic algorithm (GA), least absolute shrinkage and selection operator (LASSO), and permutation feature importance (PFI). The optimized model, utilizing second derivative, auto scaling, permutation feature importance, and support vector machine (2nd Der-AS-PFI-SVM), achieved a prediction accuracy of 93.4 %, a Kappa value of 0.876, a Brier score of 0.030, an F1 score of 0.932, and an AUC of 0.994 on an independent prediction set. Moreover, the image data (including color information and texture information) extracted from color and X-ray images were used to construct classification models and evaluate their performance. Among them, the SVM model constructed using texture information from X -ray images performed the best, and it achieved a prediction accuracy of 63.0 % on the validation set, with a Brier score of 0.181, an F1 score of 0.518, and an AUC of 0.553. By implementing mid-level fusion and high-level data fusion based on the Stacking strategy, it was found that the model employing a high-level fusion of hyperspectral spectral information and X-ray images texture information significantly outperformed the model using only hyperspectral spectral information. This advanced model attained a prediction accuracy of 95.2 %, a Kappa value of 0.912, a Brier score of 0.027, an F1 score of 0.952, and an AUC of 0.997 on the independent prediction set. In summary, this study not only provides a novel technical path for fast and non-destructive traceability of Panax ginseng origin, but also demonstrates the great potential of the combined application of HSI and X-ray technology in the field of traceability of both medicinal and food products.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    自动驾驶是一个不断发展的研究领域,为科学带来了好处,经济,和社会。虽然这方面有很多研究,目前还没有完全自主的车辆,特别是,用于越野导航。自动驾驶汽车(AV)导航是一个复杂的过程,基于多种技术和算法的数据采集的应用,管理和理解。特别是,自动驾驶辅助系统支持传感和地形感知等关键功能,实时车辆映射和定位,路径预测和驱动,通信和安全措施,在其他人中。
    在这项工作中,提出了一种结合视频帧语义分割和后续实时路线规划的越野环境中车辆自动驾驶的原始方法。为了检查提案的相关性,设计了一个面向资源受限设备的越野场景辅助驾驶的模块化框架。在场景感知模块中,深度神经网络用于分割从相机获得的红-绿-蓝(RGB)图像。第二可遍历性模块融合光检测和测距(LiDAR)点云与分割的结果以创建二元占用网格地图,以在自主导航期间提供场景理解。最后,最后一个模块,基于快速探索随机树(RRT)算法,预测一条路。弗赖堡森林数据集(FFD)和RELLIS-3D数据集用于评估所提出方法的性能。本文的理论贡献包括适用于越野驾驶场景的原始图像语义分割方法,以及使最短路线搜索A*和RRT算法适应AV路径规划。
    报告的结果非常有希望,与以前报告的解决方案相比,显示出几个优点。FFD的分割精度为85.9%,RELLIS-3D的分割精度为79.5%,包括最常见的语义类。与其他方法相比,所提出的方法在路径规划的计算时间方面更快。
    UNASSIGNED: Autonomous driving is a growing research area that brings benefits in science, economy, and society. Although there are several studies in this area, currently there is no a fully autonomous vehicle, particularly, for off-road navigation. Autonomous vehicle (AV) navigation is a complex process based on application of multiple technologies and algorithms for data acquisition, management and understanding. Particularly, a self-driving assistance system supports key functionalities such as sensing and terrain perception, real time vehicle mapping and localization, path prediction and actuation, communication and safety measures, among others.
    UNASSIGNED: In this work, an original approach for vehicle autonomous driving in off-road environments that combines semantic segmentation of video frames and subsequent real-time route planning is proposed. To check the relevance of the proposal, a modular framework for assistive driving in off-road scenarios oriented to resource-constrained devices has been designed. In the scene perception module, a deep neural network is used to segment Red-Green-Blue (RGB) images obtained from camera. The second traversability module fuses Light Detection And Ranging (LiDAR) point clouds with the results of segmentation to create a binary occupancy grid map to provide scene understanding during autonomous navigation. Finally, the last module, based on the Rapidly-exploring Random Tree (RRT) algorithm, predicts a path. The Freiburg Forest Dataset (FFD) and RELLIS-3D dataset were used to assess the performance of the proposed approach. The theoretical contributions of this article consist of the original approach for image semantic segmentation fitted to off-road driving scenarios, as well as adapting the shortest route searching A* and RRT algorithms to AV path planning.
    UNASSIGNED: The reported results are very promising and show several advantages compared to previously reported solutions. The segmentation precision achieves 85.9% for FFD and 79.5% for RELLIS-3D including the most frequent semantic classes. While compared to other approaches, the proposed approach is faster regarding computational time for path planning.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    提高过氧化氢(H2O2)检测在日常生产生活中的易操作性和便携性,具有重要的应用价值。然而,实现H2O2的快速比色检测和颜色变化定量仍然是一个挑战。在这项研究中,利用机器视觉实现了具有富氧空位的MoOx(2≤x≤3)纳米粒子对H2O2的快速、直观检测。随着H2O2浓度的增加,检测系统表现出从蓝色到绿色再到黄色的可见多色变化,紫外-可见分光光度计在680nm附近的吸收峰逐渐减小。具有出色的灵敏度,线性范围为0.1-600μmol/L,浓度低至0.1μmol/L可以检测到对H2O2具有良好的选择性。通过XPS等表征方法揭示了通过MoOx中氧空位的变化检测H2O2的传感机理,EPR,和DFT。此外,色调,饱和度,构建了基于MoOx的数值(HSV)可视化分析系统,便携式,以及在实际应用场景中对H2O2的敏感监测。这项工作提供了一个易于操作,低成本,便于实现H2O2的快速比色测定,在日常生活和工业生产中具有广阔的应用前景。
    Improving the ease of operation and portability of hydrogen peroxide (H2O2) detection in daily production and life holds significant application value. However, it remains a challenge to achieve rapid colorimetric detection of H2O2 and color change quantification. In this study, we achieved rapid and visual detection of H2O2 by MoOx (2 ≤ x ≤ 3) nanoparticles with rich oxygen vacancies using machine vision. As the concentration of H2O2 increases, the detection system exhibited a visible multi-color change from blue to green and then yellow and the absorption peak near 680 nm measured by the UV-visible spectrophotometer gradually decreased. With excellent sensitivity, a wide linear range of 0.1-600 μmol/L, concentrations as low as 0.1 μmol/L can be detected with good selectivity towards H2O2. The sensing mechanism of detecting H2O2 by the change of oxygen vacancies in MoOx was revealed through characterization methods such as XPS, EPR, and DFT. In addition, the Hue, Saturation, Value (HSV) visual analysis system based on MoOx was constructed to assist in the rapid, portable, and sensitive monitoring of H2O2 in practical application scenarios. This work offers an easy-to operate, low cost, and convenience for achieving rapid colorimetric determination of H2O2 and has broad application prospects in daily life and industrial production.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    机器视觉是热锻件理想的非接触测量方法,由于热锻件工作条件的多样性,图像分割在性能和鲁棒性方面一直是一个具有挑战性的问题。因此,本文针对锻造图像提出了一种高效、鲁棒的活动轮廓模型和相应的图像分割方法,通过测量锻件的几何参数,进行了验证实验,以证明分割方法的性能。具体来说,基于用于锻造图像的等效灰度表面的几何连续性,定义了三种类型的连续性参数;因此,提出了一种新的图像力和外部能量函数来形成新的主动轮廓模型,几何连续性蛇(GC蛇),这更适合于锻造图像的灰度分布特征,以鲁棒地提高主动轮廓的收敛性;此外,提出了一种GCSnakes初始控制点的生成策略,以构成一种高效,鲁棒的图像分割方法。实验结果表明,对于不同温度和大小的锻造图像,与现有的活动轮廓模型相比,提出的GCSnakes具有更好的分割性能,为热锻件的几何参数测量提供了更好的性能和效率。GCSnakes的最大定位和尺寸误差为0.5525mm和0.3868mm,分别,与Snakes模型的0.7873mm和0.6868mm的误差相比。
    Machine vision is a desirable non-contact measurement method for hot forgings, as image segmentation has been a challenging issue in performance and robustness resulting from the diversity of working conditions for hot forgings. Thus, this paper proposes an efficient and robust active contour model and corresponding image segmentation approach for forging images, by which verification experiments are conducted to prove the performance of the segmentation method by measuring geometric parameters for forging parts. Specifically, three types of continuity parameters are defined based on the geometric continuity of equivalent grayscale surfaces for forging images; hence, a new image force and external energy functional are proposed to form a new active contour model, Geometric Continuity Snakes (GC Snakes), which is more percipient to the grayscale distribution characteristics of forging images to improve the convergence for active contour robustly; additionally, a generating strategy for initial control points for GC Snakes is proposed to compose an efficient and robust image segmentation approach. The experimental results show that the proposed GC Snakes has better segmentation performance compared with existing active contour models for forging images of different temperatures and sizes, which provides better performance and efficiency in geometric parameter measurement for hot forgings. The maximum positioning and dimension errors by GC Snakes are 0.5525 mm and 0.3868 mm, respectively, compared with errors of 0.7873 mm and 0.6868 mm by the Snakes model.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    本文演示了如何使用UV/VIS成像来评估压碎强度,脆性,仅由白色成分组成的片剂的崩解时间和溶出曲线。使用不同水平的压缩力和无水咖啡因的API含量生产样品。使用UV照明从样品的两侧获取图像,用于API含量预测。而其他参数使用VIS照明评估。根据UV图像的颜色直方图,API含量预测相对误差为5.6%。VIS图像的纹理分析得出了在10%相对误差下的抗碎强度预测。关于脆性,根据样品的失重情况建立三组。同样,对崩解时间的评估导致了三组的识别:小于10s,11-35s,超过36秒。通过机器学习算法实现了样本的成功分类。最后,使用人工神经网络在5%的RMSE下准确预测了速释溶出曲线。图像采集过程中的50ms曝光时间及其结果强调了机器视觉在固体剂型中实时质量控制的实用性,无论API的颜色如何。
    The paper provides a demonstration of how UV/VIS imaging can be employed to evaluate the crushing strength, friability, disintegration time and dissolution profile of tablets comprised of solely white components. The samples were produced using different levels of compression force and API content of anhydrous caffeine. Images were acquired from both sides of the samples using UV illumination for the API content prediction, while the other parameters were assessed using VIS illumination. Based on the color histograms of the UV images, API content was predicted with 5.6 % relative error. Textural analysis of the VIS images yielded crushing strength predictions under 10 % relative error. Regarding friability, three groups were established according to the weight loss of the samples. Likewise, the evaluation of disintegration time led to the identification of three groups: less than 10 s, 11-35 s, and over 36 s. Successful classification of the samples was achieved with machine learning algorithms. Finally, immediate release dissolution profiles were accurately predicted under 5 % of RMSE with an artificial neural network. The 50 ms exposition time during image acquisition and the resulting outcomes underscore the practicality of machine vision for real-time quality control in solid dosage forms, regardless of the color of the API.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    早期识别老年人的认知障碍可以减轻与年龄相关的残疾的负担。步态参数与认知衰退相关并可预测认知衰退。尽管在认知研究中已经使用了多种传感器和机器学习分析方法,需要一种深度优化的基于机器视觉的步态分析方法来识别认知衰退.
    这项研究使用了158名名为华西医院老年人步态的成年人的步行录像数据集,在简短的便携式精神状态问卷上被标记为表现。我们提出了一种新颖的识别网络,深度优化GaitPart(DO-GaitPart),基于轮廓和骨骼步态图像。应用了三个改进:在模板生成阶段使用短期时间模板生成器(STTG),以降低计算成本并最大程度地减少时间信息的损失;深度空间特征提取器(DSFE)从步态图像中提取全局和局部细粒度空间特征;以及多尺度时间聚合(MTA),一种基于注意力机制的时间建模方法,以提高步态模式的可分辨性。
    消融测试表明,DO-GaitPart的每个组件都是必不可少的。DO-GaitPart在CASIA-B数据集上的背包行走场景中表现出色,优于比较方法,它们是GaitSet,GaitPart,MT3D,3D本地,TransGait,CSTL,GLN,Gait3D数据集上的GaitGL和SMPLGait。提出的机器视觉步态特征识别方法在认知状态分类任务上实现了0.876(0.852-0.900)的接收机工作特征/曲线下面积(ROCAUC)。
    所提出的方法从步态视频数据集中很好地识别了认知衰退,使其成为认知评估中的前瞻性原型工具。
    UNASSIGNED: Early identification of cognitive impairment in older adults could reduce the burden of age-related disabilities. Gait parameters are associated with and predictive of cognitive decline. Although a variety of sensors and machine learning analysis methods have been used in cognitive studies, a deep optimized machine vision-based method for analyzing gait to identify cognitive decline is needed.
    UNASSIGNED: This study used a walking footage dataset of 158 adults named West China Hospital Elderly Gait, which was labelled by performance on the Short Portable Mental Status Questionnaire. We proposed a novel recognition network, Deep Optimized GaitPart (DO-GaitPart), based on silhouette and skeleton gait images. Three improvements were applied: short-term temporal template generator (STTG) in the template generation stage to decrease computational cost and minimize loss of temporal information; depth-wise spatial feature extractor (DSFE) to extract both global and local fine-grained spatial features from gait images; and multi-scale temporal aggregation (MTA), a temporal modeling method based on attention mechanism, to improve the distinguishability of gait patterns.
    UNASSIGNED: An ablation test showed that each component of DO-GaitPart was essential. DO-GaitPart excels in backpack walking scene on CASIA-B dataset, outperforming comparison methods, which were GaitSet, GaitPart, MT3D, 3D Local, TransGait, CSTL, GLN, GaitGL and SMPLGait on Gait3D dataset. The proposed machine vision gait feature identification method achieved a receiver operating characteristic/area under the curve (ROCAUC) of 0.876 (0.852-0.900) on the cognitive state classification task.
    UNASSIGNED: The proposed method performed well identifying cognitive decline from the gait video datasets, making it a prospective prototype tool in cognitive assessment.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    高效,无损收获白菜对于保持其风味和质量至关重要。目前白菜收获主要依靠机械化自动采摘方法。然而,大多数现有的卷心菜收获装置的一个显著缺陷是没有一个根部姿势识别系统来及时调整根部姿势,因此影响采收过程中切根的准确性。为了解决这个问题,本研究介绍了一种将深度学习与传统图像处理算法相结合的白菜根姿态识别方法。通过YOLOv5s深度学习模型对白菜的主根感兴趣区域(ROI)区域进行初步检测。随后,传统的图像处理方法,格雷厄姆算法,并采用计算最小外接矩形的方法来具体检测白菜根的倾角。这种方法有效地解决了计算由外叶遮挡引起的根倾角的困难。结果表明,该方法的准确率和召回率分别为98.7%和98.6%。分别,姿态的平均绝对误差为0.80°,平均相对误差为1.34%。将该方法作为机械收获的参考,可以有效减轻白菜的伤害率。
    Efficient, non-destructive cabbage harvesting is crucial for preserving its flavor and quality. Current cabbage harvesting mainly relies on mechanized automatic picking methods. However, a notable deficiency in most existing cabbage harvesting devices is the absence of a root posture recognition system to promptly adjust the root posture, consequently impacting the accuracy of root cutting during harvesting. To address this issue, this study introduces a cabbage root posture recognition method that combines deep learning with traditional image processing algorithms. Preliminary detection of the main root Region of Interest (ROI) areas of the cabbage is carried out through the YOLOv5s deep learning model. Subsequently, traditional image processing methods, the Graham algorithm, and the method of calculating the minimum circumscribed rectangle are employed to specifically detect the inclination angle of cabbage roots. This approach effectively addresses the difficulty in calculating the inclination angle of roots caused by occlusion from outer leaves. The results demonstrate that the precision and recall of this method are 98.7 % and 98.6 %, respectively, with an average absolute error of 0.80° and an average relative error of 1.34 % in posture. Using this method as a reference for mechanical harvesting can effectively mitigate cabbage damage rates.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    目前,利用AI图像识别技术对海量电网输电线路巡检图片进行智能缺陷检测是一种高效、流行的方法。通常,缺陷检测算法模型的构建有两条技术路线:一是使用轻量级网络,这提高了效率,但它通常只能针对少数类型的缺陷,并可能降低检测精度;另一种是使用复杂的网络模型,这提高了准确性,并且可以同时识别多种类型的缺陷,但是它的计算量大,效率低。为了保持模型的高检测精度以及其轻量化结构,提出了一种轻量高效的基于DCP-YOLOv8的输电线路多类型缺陷检测方法。该方法采用可变形卷积(C2f_DCNv3)来增强缺陷特征提取能力,并设计了一种重新参数化的交叉相位特征融合结构(RCSP),将高层语义特征与低层空间特征进行优化和融合,从而提高模型识别不同尺度缺陷的能力,同时显著降低模型参数;它结合了动态检测头和可变形卷积v3的检测头(DCNv3-Dyhead),以增强特征表达能力和上下文信息的利用率,进一步提高检测精度。实验结果表明,在一个包含20条真实输电线路缺陷的数据集上,该方法将平均准确度(mAP@0.5)提高到72.2%,增长4.3%,与最轻的基线YOLOv8n模型相比;模型参数数量仅为2.8M,减少9.15%,每秒处理帧数(FPS)达到103,满足实时检测需求。在多类型缺陷检测的场景中,它有效地平衡了检测精度和性能与定量泛化。
    Currently, the intelligent defect detection of massive grid transmission line inspection pictures using AI image recognition technology is an efficient and popular method. Usually, there are two technical routes for the construction of defect detection algorithm models: one is to use a lightweight network, which improves the efficiency, but it can generally only target a few types of defects and may reduce the detection accuracy; the other is to use a complex network model, which improves the accuracy, and can identify multiple types of defects at the same time, but it has a large computational volume and low efficiency. To maintain the model\'s high detection accuracy as well as its lightweight structure, this paper proposes a lightweight and efficient multi type defect detection method for transmission lines based on DCP-YOLOv8. The method employs deformable convolution (C2f_DCNv3) to enhance the defect feature extraction capability, and designs a re-parameterized cross phase feature fusion structure (RCSP) to optimize and fuse high-level semantic features with low level spatial features, thus improving the capability of the model to recognize defects at different scales while significantly reducing the model parameters; additionally, it combines the dynamic detection head and deformable convolutional v3\'s detection head (DCNv3-Dyhead) to enhance the feature expression capability and the utilization of contextual information to further improve the detection accuracy. Experimental results show that on a dataset containing 20 real transmission line defects, the method increases the average accuracy (mAP@0.5) to 72.2%, an increase of 4.3%, compared with the lightest baseline YOLOv8n model; the number of model parameters is only 2.8 M, a reduction of 9.15%, and the number of processed frames per second (FPS) reaches 103, which meets the real time detection demand. In the scenario of multi type defect detection, it effectively balances detection accuracy and performance with quantitative generalizability.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    同时定位和映射(SLAM)是大多数自治系统的关键功能,允许他们浏览并创建不熟悉环境的地图。传统的视觉SLAM,也通常被称为VSLAM,依赖于基于框架的相机和结构化的处理管道,在动态或弱光环境中面临挑战。然而,事件相机技术和神经形态处理的最新进展为克服这些限制提供了有希望的机会。受生物视觉系统启发的事件摄像机异步捕获场景,消耗最小的功率,但具有更高的时间分辨率。神经形态处理器,它们旨在模仿人脑的并行处理能力,为基于事件的数据流的实时数据处理提供有效的计算。本文全面概述了将事件摄像机和神经形态处理器集成到VSLAM系统中的最新研究工作。它讨论了事件摄像机和神经形态处理器背后的原理,突出了它们相对于传统传感和处理方法的优势。此外,对基于事件的SLAM中的最新方法进行了深入的调查,包括特征提取,运动估计,和地图重建技术。此外,事件摄像机与神经形态处理器的集成,专注于它们在能源效率方面的协同优势,鲁棒性,和实时性能,被探索了。本文还讨论了这一新兴领域的挑战和开放的研究问题,如传感器校准,数据融合,和算法开发。最后,概述了基于事件的SLAM系统的潜在应用和未来方向,从机器人和自动驾驶汽车到增强现实。
    Simultaneous Localization and Mapping (SLAM) is a crucial function for most autonomous systems, allowing them to both navigate through and create maps of unfamiliar surroundings. Traditional Visual SLAM, also commonly known as VSLAM, relies on frame-based cameras and structured processing pipelines, which face challenges in dynamic or low-light environments. However, recent advancements in event camera technology and neuromorphic processing offer promising opportunities to overcome these limitations. Event cameras inspired by biological vision systems capture the scenes asynchronously, consuming minimal power but with higher temporal resolution. Neuromorphic processors, which are designed to mimic the parallel processing capabilities of the human brain, offer efficient computation for real-time data processing of event-based data streams. This paper provides a comprehensive overview of recent research efforts in integrating event cameras and neuromorphic processors into VSLAM systems. It discusses the principles behind event cameras and neuromorphic processors, highlighting their advantages over traditional sensing and processing methods. Furthermore, an in-depth survey was conducted on state-of-the-art approaches in event-based SLAM, including feature extraction, motion estimation, and map reconstruction techniques. Additionally, the integration of event cameras with neuromorphic processors, focusing on their synergistic benefits in terms of energy efficiency, robustness, and real-time performance, was explored. The paper also discusses the challenges and open research questions in this emerging field, such as sensor calibration, data fusion, and algorithmic development. Finally, the potential applications and future directions for event-based SLAM systems are outlined, ranging from robotics and autonomous vehicles to augmented reality.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    微塑料(MPs,尺寸≤5毫米)已成为全球关注的重要问题,威胁海洋和淡水生态系统,MP检测技术的缺乏是值得注意的。这项研究的主要目标是开发一种摄像机传感器,用于检测MP并在运动中测量其大小和速度。这项研究介绍了一种涉及计算机视觉和人工智能(AI)的新型方法,用于检测MP。三种不同的摄像系统,包括固定焦点2D和自动对焦(2D和3D),进行了实施和比较。基于YOLOv5的对象检测模型用于检测捕获图像中的MP。然后实施了DeepSORT,用于通过连续图像跟踪MP。在实验室水槽环境中进行实时测试,MP计数的精度为97%,在当地河流的现场测试中,精确度为96%。这项研究提供了利用人工智能在不同环境环境中检测MP的基础见解,有助于更有效的努力和战略来管理和减轻MP污染。
    Microplastics (MPs, size ≤ 5 mm) have emerged as a significant worldwide concern, threatening marine and freshwater ecosystems, and the lack of MP detection technologies is notable. The main goal of this research is the development of a camera sensor for the detection of MPs and measuring their size and velocity while in motion. This study introduces a novel methodology involving computer vision and artificial intelligence (AI) for the detection of MPs. Three different camera systems, including fixed-focus 2D and autofocus (2D and 3D), were implemented and compared. A YOLOv5-based object detection model was used to detect MPs in the captured image. DeepSORT was then implemented for tracking MPs through consecutive images. In real-time testing in a laboratory flume setting, the precision in MP counting was found to be 97%, and during field testing in a local river, the precision was 96%. This study provides foundational insights into utilizing AI for detecting MPs in different environmental settings, contributing to more effective efforts and strategies for managing and mitigating MP pollution.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

公众号