3D point clouds

3D 点云
  • 文章类型: Journal Article
    输电线路走廊点云场景中目标对象的语义分割是电力线树屏障检测的关键步骤。大量的,无序分布,输电线路走廊场景中点云的非均匀性对特征提取提出了重大挑战。以往的研究往往忽视了空间信息的核心利用,限制了网络理解复杂几何形状的能力。为了克服这个限制,本文着眼于增强分割网络中空间几何信息的深度表达,并提出了一种称为BDF-Net的方法来改进RandLA-Net。对于每个输入的3D点云数据,BDF-Net首先通过空间信息编码块将相对坐标和相对距离信息编码为空间几何特征表示,以捕获点云数据的局部空间结构。随后,双线性池块通过利用其双线性相互作用能力有效地将点云的特征信息与空间几何表示相结合,从而学习更多的区别性局部特征描述符。全局特征提取块利用点位置与相对位置的比值捕获点云数据中的全局结构信息,从而增强网络的语义理解能力。为了验证BDF-Net的性能,本文构建了一个数据集,PPCD,针对输电线路走廊的点云场景进行了详细的实验。实验结果表明,BDF-Net在各种评估指标上实现了显著的性能提升,具体实现97.16%的OA,77.48%的mIoU,mAcc为87.6%,为3.03%,16.23%,比RandLA-Net高18.44%,分别。此外,与其他最新方法的比较也验证了BDF-Net在点云语义分割任务中的优越性。
    Semantic segmentation of target objects in power transmission line corridor point cloud scenes is a crucial step in powerline tree barrier detection. The massive quantity, disordered distribution, and non-uniformity of point clouds in power transmission line corridor scenes pose significant challenges for feature extraction. Previous studies have often overlooked the core utilization of spatial information, limiting the network\'s ability to understand complex geometric shapes. To overcome this limitation, this paper focuses on enhancing the deep expression of spatial geometric information in segmentation networks and proposes a method called BDF-Net to improve RandLA-Net. For each input 3D point cloud data, BDF-Net first encodes the relative coordinates and relative distance information into spatial geometric feature representations through the Spatial Information Encoding block to capture the local spatial structure of the point cloud data. Subsequently, the Bilinear Pooling block effectively combines the feature information of the point cloud with the spatial geometric representation by leveraging its bilinear interaction capability thus learning more discriminative local feature descriptors. The Global Feature Extraction block captures the global structure information in the point cloud data by using the ratio between the point position and the relative position, so as to enhance the semantic understanding ability of the network. In order to verify the performance of BDF-Net, this paper constructs a dataset, PPCD, for the point cloud scenario of transmission line corridors and conducts detailed experiments on it. The experimental results show that BDF-Net achieves significant performance improvements in various evaluation metrics, specifically achieving an OA of 97.16%, a mIoU of 77.48%, and a mAcc of 87.6%, which are 3.03%, 16.23%, and 18.44% higher than RandLA-Net, respectively. Moreover, comparisons with other state-of-the-art methods also verify the superiority of BDF-Net in point cloud semantic segmentation tasks.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    本文提出了一种新颖的分割算法,专门为具有高变异性和噪声的3D点云应用而开发,特别适用于文物建筑的三维数据。该方法可以在基于边缘检测的分割过程中进行分类。此外,它使用从3D点云的超体素化生成的基于图形的拓扑结构,用于使边缘点闭合并定义不同的段。该算法为生成结果提供了有价值的工具,这些结果可用于后续的分类任务和处理3D点云的更广泛的计算机应用。这种分割方法的特点之一是它是无监督的,这使得它对于标记数据稀缺的传统应用特别有利。它也很容易适应不同的边缘点检测和超体素化算法。最后,结果表明,三维数据可以分割成不同的建筑元素,这对于进一步分类或识别很重要。对历史建筑的真实数据进行的大量测试证明了该方法的有效性。结果表明,与其他三种分割方法相比,性能更优越,无论是在全球范围内还是在历史建筑的平面和弯曲区域的分割中。
    This paper presents a novel segmentation algorithm specially developed for applications in 3D point clouds with high variability and noise, particularly suitable for heritage building 3D data. The method can be categorized within the segmentation procedures based on edge detection. In addition, it uses a graph-based topological structure generated from the supervoxelization of the 3D point clouds, which is used to make the closure of the edge points and to define the different segments. The algorithm provides a valuable tool for generating results that can be used in subsequent classification tasks and broader computer applications dealing with 3D point clouds. One of the characteristics of this segmentation method is that it is unsupervised, which makes it particularly advantageous for heritage applications where labelled data is scarce. It is also easily adaptable to different edge point detection and supervoxelization algorithms. Finally, the results show that the 3D data can be segmented into different architectural elements, which is important for further classification or recognition. Extensive testing on real data from historic buildings demonstrated the effectiveness of the method. The results show superior performance compared to three other segmentation methods, both globally and in the segmentation of planar and curved zones of historic buildings.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    本文提出了一种自主机器人系统,使用深度学习的语义分割和关节式机械手来修剪甜椒叶。该系统涉及三个主要任务:作物部分的感知,修剪位置的检测,以及关节式机械手的控制。利用语义分割神经网络对甜椒植株的不同部位进行识别,然后用于创建3D点云,以检测修剪位置和操纵器姿势。最终,控制机械手机器人修剪作物部分。本文详细介绍了构建甜椒修剪系统所涉及的三个任务以及如何将它们集成在一起。在实验中,我们使用机械臂在一定高度范围内操纵修剪叶片的动作,并使用深度相机获得3D点云。控制程序是使用在ROS(机器人操作系统)上运行的各种编程语言在不同的模块中开发的。
    This paper proposes an autonomous robotic system to prune sweet pepper leaves using semantic segmentation with deep learning and an articulated manipulator. This system involves three main tasks: the perception of crop parts, the detection of pruning position, and the control of the articulated manipulator. A semantic segmentation neural network is employed to recognize the different parts of the sweet pepper plant, which is then used to create 3D point clouds for detecting the pruning position and the manipulator pose. Eventually, a manipulator robot is controlled to prune the crop part. This article provides a detailed description of the three tasks involved in building the sweet pepper pruning system and how to integrate them. In the experiments, we used a robot arm to manipulate the pruning leaf actions within a certain height range and a depth camera to obtain 3D point clouds. The control program was developed in different modules using various programming languages running on the ROS (Robot Operating System).
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    近年来,高光谱成像技术已广泛应用于农业中,以评估叶片水分含量等复杂的植物生理性状,营养水平,和疾病压力。该技术的关键组成部分是白色参考,用于消除不同波长的非均匀照明强度对原始高光谱图像的影响。然而,平坦的白色瓷砖不能准确反映植物叶片上的照明强度变化,由于叶几何形状(例如,倾斜角)及其与光照的相互作用严重影响植物的反射光谱和植被指数,例如归一化植被指数(NDVI)。在这项研究中,综述了叶片角度对植物反射光谱的影响,建立了叶片高光谱图像与三维点云融合的图像标定模型。使用室内台式高光谱成像系统以不同的倾斜角和方向对玉米和大豆叶样品进行成像,并分析NDVI值的差异。结果表明,叶片的NDVI随角度有很大变化。两个物种之间随角度的变化趋势有所不同。使用从与高光谱图像同时拍摄的3D点云数据获得的叶片倾斜角和方向的测量,成功开发了支持向量回归(SVR)模型,以将叶片上不同角度的像素的NDVI值校准为相同的标准,就像将叶片平放在水平面上一样。玉米和大豆的测量和预测的叶片角度影响之间的R平方值分别为0.76和0.94,分别。该方法具有用于任何一般植物成像系统以改善表型质量的潜力。
    During recent years, hyperspectral imaging technologies have been widely applied in agriculture to evaluate complex plant physiological traits such as leaf moisture content, nutrient level, and disease stress. A critical component of this technique is white referencing used to remove the effect of non-uniform lighting intensity in different wavelengths on raw hyperspectral images. However, a flat white tile cannot accurately reflect the lighting intensity variance on plant leaves, since the leaf geometry (e.g., tilt angles) and its interaction with the illumination severely impact plant reflectance spectra and vegetation indices such as the normalized difference vegetation index (NDVI). In this research, the impacts of leaf angles on plant reflectance spectra were summarized, and an improved image calibration model using the fusion of leaf hyperspectral images and 3D point clouds was built. Corn and soybean leaf samples were imaged at different tilt angles and orientations using an indoor desktop hyperspectral imaging system and analyzed for differences in the NDVI values. The results showed that the leaf\'s NDVI largely changed with angles. The changing trends with angles differed between the two species. Using measurements of leaf tilt angle and orientation obtained from the 3D point cloud data taken simultaneously with the hyperspectral images, a support vector regression (SVR) model was successfully developed to calibrate the NDVI values of pixels at different angles on a leaf to a same standard as if the leaf was laid flat on a horizontal surface. The R-squared values between the measured and predicted leaf angle impacts were 0.76 and 0.94 for corn and soybean, respectively. This method has a potential to be used in any general plant imaging systems to improve the phenotyping quality.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    本文提出了一种深度学习框架,用于编码面部和骨骼形状之间的特定主题转换,以进行正颌手术计划。我们的框架涉及双向点对点卷积网络(P2P-Conv),以预测面部和骨骼形状之间的转换。P2P-Conv是最先进的P2P-Net的扩展,并利用动态逐点卷积(即,PointConv)来捕获本地到全球的空间信息。数据增强是在P2P-Conv的训练中使用来自面部和骨骼形状的多个点子集进行的。在推理过程中,为多个点子集生成的网络输出被组合成密集变换。最后,使用相干点漂移(CPD)算法的非刚性配准用于基于预测的点集生成表面网格。对真实受试者数据的实验结果表明,与最先进的方法相比,我们的方法大大提高了对面部和骨骼形状的预测。
    This paper proposes a deep learning framework to encode subject-specific transformations between facial and bony shapes for orthognathic surgical planning. Our framework involves a bidirectional point-to-point convolutional network (P2P-Conv) to predict the transformations between facial and bony shapes. P2P-Conv is an extension of the state-of-the-art P2P-Net and leverages dynamic point-wise convolution (i.e., PointConv) to capture local-to-global spatial information. Data augmentation is carried out in the training of P2P-Conv with multiple point subsets from the facial and bony shapes. During inference, network outputs generated for multiple point subsets are combined into a dense transformation. Finally, non-rigid registration using the coherent point drift (CPD) algorithm is applied to generate surface meshes based on the predicted point sets. Experimental results on real-subject data demonstrate that our method substantially improves the prediction of facial and bony shapes over state-of-the-art methods.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    我们提出3DPointCaps++用于学习健壮,灵活和可概括的3D对象表示,而不需要繁重的注释工作或监督。与传统的3D生成模型不同,我们的算法旨在建立一个结构化的潜在空间,其中某些因素的形状变化,例如对象部分,可以解开成独立的子空间。然后,我们的新颖解码器使用去卷积算子以自我监督的方式对这些单独的潜在子空间(即胶囊)进行操作,以重建3D点。我们进一步引入了聚类损失,以确保由单个胶囊重建的点保持局部,并且不会无法控制地散布在对象上。这些贡献使我们的网络能够解决零件分割的挑战性任务,零件插值/替换以及刚性/非刚性形状的对应估计,和跨/内类别。我们对ShapeNet对象和人工扫描的广泛评估表明,我们的网络可以学习在许多应用中健壮和有用的通用表示。
    We present 3DPointCaps++ for learning robust, flexible and generalizable 3D object representations without requiring heavy annotation efforts or supervision. Unlike conventional 3D generative models, our algorithm aims for building a structured latent space where certain factors of shape variations, such as object parts, can be disentangled into independent sub-spaces. Our novel decoder then acts on these individual latent sub-spaces (i.e. capsules) using deconvolution operators to reconstruct 3D points in a self-supervised manner. We further introduce a cluster loss ensuring that the points reconstructed by a single capsule remain local and do not spread across the object uncontrollably. These contributions allow our network to tackle the challenging tasks of part segmentation, part interpolation/replacement as well as correspondence estimation across rigid / non-rigid shape, and across / within category. Our extensive evaluations on ShapeNet objects and human scans demonstrate that our network can learn generic representations that are robust and useful in many applications.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    在过去的几十年里,消费级RGB-D(红绿蓝深度)相机已经在农业环境中的多种应用中获得了普及。有趣的是,这些相机用于空间映射,可以用于机器人定位和导航。为农业领域的目标机器人应用绘制环境图是一项特别具有挑战性的任务,由于高度的时空变异性,可能的不利光照条件,以及这些环境的不可预测性。本研究的目的是研究使用RGB-D相机和无人地面车辆(UGV)自动绘制商业果园的环境图,并提供有关树木高度和冠层体积的信息。将地面制图系统的结果与无人机(UAV)获得的三维(3D)正交osaics进行了比较。总的来说,两种传感方法都导致了类似的高度测量,虽然RGB-D相机更准确地计算了树木的体积,因为地面系统捕获的3D点云更加详细。最后,两个数据集的融合提供了树的最精确表示。
    During the last decades, consumer-grade RGB-D (red green blue-depth) cameras have gained popularity for several applications in agricultural environments. Interestingly, these cameras are used for spatial mapping that can serve for robot localization and navigation. Mapping the environment for targeted robotic applications in agricultural fields is a particularly challenging task, owing to the high spatial and temporal variability, the possible unfavorable light conditions, and the unpredictable nature of these environments. The aim of the present study was to investigate the use of RGB-D cameras and unmanned ground vehicle (UGV) for autonomously mapping the environment of commercial orchards as well as providing information about the tree height and canopy volume. The results from the ground-based mapping system were compared with the three-dimensional (3D) orthomosaics acquired by an unmanned aerial vehicle (UAV). Overall, both sensing methods led to similar height measurements, while the tree volume was more accurately calculated by RGB-D cameras, as the 3D point cloud captured by the ground system was far more detailed. Finally, fusion of the two datasets provided the most precise representation of the trees.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    这项研究测试了机器学习(ML)方法是否可以有效地将单个植物从复杂的3D冠层激光扫描中分离出来,作为分析特定植物特征的先决条件。为此,我们用PlantEye(R)激光扫描仪扫描绿豆和鹰嘴豆作物。首先,我们使用区域生长分割算法从3D空间中的背景分割作物冠层。然后,对基于卷积神经网络(CNN)的ML算法进行了微调,以进行植物计数。只有在我们将数据的维数降低到2D之后,才有可能应用基于CNN(卷积神经网络)的处理架构。这允许识别单个植物及其计数,绿豆和鹰嘴豆植物的准确率为93.18%和92.87%,分别。这些步骤与表型管道相连,它现在可以取代低效的手动计数操作,昂贵的,而且容易出错。CNN在这项研究中的使用通过降维来创新性地解决,添加高度信息作为颜色,以及基于二维CNN的方法的后续应用。我们发现在3D信息上使用ML存在很大差距。这个差距必须解决,特别是对于更复杂的植物特征提取,我们打算通过进一步的研究来实施。
    This study tested whether machine learning (ML) methods can effectively separate individual plants from complex 3D canopy laser scans as a prerequisite to analyzing particular plant features. For this, we scanned mung bean and chickpea crops with PlantEye (R) laser scanners. Firstly, we segmented the crop canopies from the background in 3D space using the Region Growing Segmentation algorithm. Then, Convolutional Neural Network (CNN) based ML algorithms were fine-tuned for plant counting. Application of the CNN-based (Convolutional Neural Network) processing architecture was possible only after we reduced the dimensionality of the data to 2D. This allowed for the identification of individual plants and their counting with an accuracy of 93.18% and 92.87% for mung bean and chickpea plants, respectively. These steps were connected to the phenotyping pipeline, which can now replace manual counting operations that are inefficient, costly, and error-prone. The use of CNN in this study was innovatively solved with dimensionality reduction, addition of height information as color, and consequent application of a 2D CNN-based approach. We found there to be a wide gap in the use of ML on 3D information. This gap will have to be addressed, especially for more complex plant feature extractions, which we intend to implement through further research.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    In the current industrial landscape, increasingly pervaded by technological innovations, the adoption of optimized strategies for asset management is becoming a critical key success factor. Among the various strategies available, the \"Prognostics and Health Management\" strategy is able to support maintenance management decisions more accurately, through continuous monitoring of equipment health and \"Remaining Useful Life\" forecasting. In the present study, convolutional neural network-based deep neural network techniques are investigated for the remaining useful life prediction of a punch tool, whose degradation is caused by working surface deformations during the machining process. Surface deformation is determined using a 3D scanning sensor capable of returning point clouds with micrometric accuracy during the operation of the punching machine, avoiding both downtime and human intervention. The 3D point clouds thus obtained are transformed into bidimensional image-type maps, i.e., maps of depths and normal vectors, to fully exploit the potential of convolutional neural networks for extracting features. Such maps are then processed by comparing 15 genetically optimized architectures with the transfer learning of 19 pretrained models, using a classic machine learning approach, i.e., support vector regression, as a benchmark. The achieved results clearly show that, in this specific case, optimized architectures provide performance far superior (MAPE = 0.058) to that of transfer learning, which, instead, remains at a lower or slightly higher level (MAPE = 0.416) than support vector regression (MAPE = 0.857).
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    Leaf angle and leaf area index together influence canopy light interception and canopy photosynthesis. However, so far, there is no effective method to identify the optimal combination of these two parameters for canopy photosynthesis. In this study, first a robust high-throughput method for accurate segmentation of maize organs based on 3D point clouds data was developed, then the segmented plant organs were used to generate new 3D point clouds for the canopy of altered architectures. With this, we simulated the synergistic effect of leaf area and leaf angle on canopy photosynthesis. The results show that, compared to the traditional parameters describing the canopy photosynthesis including leaf area index, facet angle and canopy coverage, a new parameter - the canopy occupation volume (COV) - can better explain the variations of canopy photosynthetic capacity. Specifically, COV can explain > 79% variations of canopy photosynthesis generated by changing leaf angle and > 84% variations of canopy photosynthesis generated by changing leaf area. As COV can be calculated in a high-throughput manner based on the canopy point clouds, it can be used to evaluate canopy architecture in breeding and agronomic research.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Sci-hub)

公众号