realsense

  • 文章类型: Journal Article
    技术进步扩大了捕获人体运动的方法范围,包括涉及惯性传感器(IMU)和光学替代品的解决方案。然而,与商业解决方案相关的复杂性和成本不断上升,促使人们探索更具成本效益的替代方案。本文介绍了一种使用RealSense深度相机和智能计算机视觉算法的无标记光学运动捕获系统。它有助于精确的姿势评估,关节角度的实时计算,以及获取特定受试者的人体测量数据以进行步态分析。与复杂的商业解决方案相比,所提出的系统以其简单性和可负担性而著称。收集的数据存储在逗号分隔值(CSV)文件中,简化后续分析和数据挖掘。初步试验,在受控的实验室环境中进行,并采用商用MEMS-IMU系统作为参考,人体测量的最大相对误差为7.6%,平均高度最大绝对误差为4.67cm。步幅长度测量显示最大相对误差为11.2%。静态接头角度试验的最大平均误差为10.2%,而动态接头角度测试显示最大平均误差为9.06%。所提出的光学系统为康复等领域的潜在应用提供了足够的精度,体育分析,和娱乐。
    Technological advancements have expanded the range of methods for capturing human body motion, including solutions involving inertial sensors (IMUs) and optical alternatives. However, the rising complexity and costs associated with commercial solutions have prompted the exploration of more cost-effective alternatives. This paper presents a markerless optical motion capture system using a RealSense depth camera and intelligent computer vision algorithms. It facilitates precise posture assessment, the real-time calculation of joint angles, and acquisition of subject-specific anthropometric data for gait analysis. The proposed system stands out for its simplicity and affordability in comparison to complex commercial solutions. The gathered data are stored in comma-separated value (CSV) files, simplifying subsequent analysis and data mining. Preliminary tests, conducted in controlled laboratory environments and employing a commercial MEMS-IMU system as a reference, revealed a maximum relative error of 7.6% in anthropometric measurements, with a maximum absolute error of 4.67 cm at average height. Stride length measurements showed a maximum relative error of 11.2%. Static joint angle tests had a maximum average error of 10.2%, while dynamic joint angle tests showed a maximum average error of 9.06%. The proposed optical system offers sufficient accuracy for potential application in areas such as rehabilitation, sports analysis, and entertainment.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    头部姿势评估可以揭示有关人体运动控制的重要临床信息。定量评估有可能客观地评估头部姿势和动作的细节,以监测疾病的进展或治疗的有效性。基于光电摄像机的运动捕捉系统,被公认为临床生物力学的黄金标准,已经被提出用于头部姿势估计。然而,这些系统需要将标记物放置在人的脸上,这对于日常临床实践是不切实际的。此外,对这类设备的有限访问以及在自然环境中评估移动性的新兴趋势支持了能够使用现成传感器估计头部方向的算法的开发,例如RGB相机。虽然人工视觉是一个热门的研究领域,基于适用于临床应用的图像识别的人体姿态估计的有限验证。本文首先简要介绍了文献中可用的头部姿态估计算法。当前最先进的头部姿势算法,旨在从视频中捕获面部几何形状,然后进一步评估和比较OpenFace2.0、MediaPipe和3DDFA_V2。通过将两种方法与基线进行比较来评估准确性,用基于光电相机的运动捕捉系统测量。结果表明,根据运动平面的不同,3DDFA_V2的平均误差小于或等于5.6。而OpenFace2.0和MediaPipe的平均误差达到14.1和11.0。分别。这证明了3DDFA_V2算法在估计头部姿势方面的优越性,在不同的运动方向,并表明该算法可用于临床场景。
    Head pose assessment can reveal important clinical information on human motor control. Quantitative assessment have the potential to objectively evaluate head pose and movements\' specifics, in order to monitor the progression of a disease or the effectiveness of a treatment. Optoelectronic camera-based motion-capture systems, recognized as a gold standard in clinical biomechanics, have been proposed for head pose estimation. However, these systems require markers to be positioned on the person\'s face which is impractical for everyday clinical practice. Furthermore, the limited access to this type of equipment and the emerging trend to assess mobility in natural environments support the development of algorithms capable of estimating head orientation using off-the-shelf sensors, such as RGB cameras. Although artificial vision is a popular field of research, limited validation of human pose estimation based on image recognition suitable for clinical applications has been performed. This paper first provides a brief review of available head pose estimation algorithms in the literature. Current state-of-the-art head pose algorithms designed to capture the facial geometry from videos, OpenFace 2.0, MediaPipe and 3DDFA_V2, are then further evaluated and compared. Accuracy is assessed by comparing both approaches to a baseline, measured with an optoelectronic camera-based motion-capture system. Results reveal a mean error lower or equal to 5.6∘ for 3DDFA_V2 depending on the plane of movement, while the mean error reaches 14.1∘ and 11.0∘ for OpenFace 2.0 and MediaPipe, respectively. This demonstrates the superiority of the 3DDFA_V2 algorithm in estimating head pose, in different directions of motion, and suggests that this algorithm can be used in clinical scenarios.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    Fruit recognition based on depth information has been a hot topic due to its advantages. However, the present equipment and methods cannot meet the requirements of rapid and reliable recognition and location of fruits in close shot for robot harvesting. To solve this problem, we propose a recognition algorithm for citrus fruit based on RealSense. This method effectively utilizes depth-point cloud data in a close-shot range of 160 mm and different geometric features of the fruit and leaf to recognize fruits with a intersection curve cut by the depth-sphere. Experiments with close-shot recognition of six varieties of fruit under different conditions were carried out. The detection rates of little occlusion and adhesion were from 80⁻100%. However, severe occlusion and adhesion still have a great influence on the overall success rate of on-branch fruits recognition, the rate being 63.8%. The size of the fruit has a more noticeable impact on the success rate of detection. Moreover, due to close-shot near-infrared detection, there was no obvious difference in recognition between bright and dark conditions. The advantages of close-shot limited target detection with RealSense, fast foreground and background removal and the simplicity of the algorithm with high precision may contribute to high real-time vision-servo operations of harvesting robots.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Sci-hub)

       PDF(Pubmed)

  • 文章类型: Journal Article
    The introduction of RGB-Depth (RGB-D) sensors into the visually impaired people (VIP)-assisting area has stirred great interest of many researchers. However, the detection range of RGB-D sensors is limited by narrow depth field angle and sparse depth map in the distance, which hampers broader and longer traversability awareness. This paper proposes an effective approach to expand the detection of traversable area based on a RGB-D sensor, the Intel RealSense R200, which is compatible with both indoor and outdoor environments. The depth image of RealSense is enhanced with IR image large-scale matching and RGB image-guided filtering. Traversable area is obtained with RANdom SAmple Consensus (RANSAC) segmentation and surface normal vector estimation, preliminarily. A seeded growing region algorithm, combining the depth image and RGB image, enlarges the preliminary traversable area greatly. This is critical not only for avoiding close obstacles, but also for allowing superior path planning on navigation. The proposed approach has been tested on a score of indoor and outdoor scenarios. Moreover, the approach has been integrated into an assistance system, which consists of a wearable prototype and an audio interface. Furthermore, the presented approach has been proved to be useful and reliable by a field test with eight visually impaired volunteers.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

公众号