image motion analysis

图像运动分析
  • 文章类型: Journal Article
    本文的目标是双重的:首先,为了提供一种新颖的数学模型,该模型基于具有四个自由度的拉格朗日力学来描述人类手指的运动链,其次,使用身体健全的个体的数据估计模型参数。在文献中,已经开发了多种数学模型来描述人类手指的运动。这些模型几乎没有提供有关潜在机制或相应运动方程的信息。此外,这些模型没有提供关于它们如何用不同的人体测量进行缩放的信息。这里使用的数据是使用实验程序生成的,该实验程序考虑每个手指段的自由响应运动以及经由运动捕获系统捕获的数据。然后对收集的角度数据进行滤波并拟合到运动方程的线性二阶微分近似。研究结果表明,节段的自由响应运动在屈曲/伸展和ad/外展上的阻尼不足。
    The goal of this paper is twofold: firstly, to provide a novel mathematical model that describes the kinematic chain of motion of the human fingers based on Lagrangian mechanics with four degrees of freedom and secondly, to estimate the model parameters using data from able-bodied individuals. In the literature there are a variety of mathematical models that have been developed to describe the motion of the human finger. These models offer little to no information on the underlying mechanisms or corresponding equations of motion. Furthermore, these models do not provide information as to how they scale with different anthropometries. The data used here is generated using an experimental procedure that considers the free response motion of each finger segment with data captured via a motion capture system. The angular data collected are then filtered and fitted to a linear second-order differential approximation of the equations of motion. The results of the study show that the free response motion of the segments is underdamped across flexion/extension and ad/abduction.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    当前的相机陷阱使用被动红外触发器;因此,它们仅在动物的表面温度与背景温度大不相同时才捕获图像。吸热动物,比如哺乳动物和鸟类,为触发摄像机提供足够的温度对比,而放热的动物,比如两栖动物,爬行动物,和无脊椎动物,不要。因此,能够监测体温变化的相机陷阱可以扩大对体温变化的动物的生态研究能力。这项研究提出了设计,发展,以及评估太阳能供电和人工智能辅助的相机陷阱系统,该系统能够监测吸热和放热动物。该系统是使用中央处理单元开发的,集成图形处理单元,摄像头,红外光,闪存驱动器,印刷电路板,太阳能电池板,电池,麦克风,GPS接收器,温度/湿度传感器,光传感器,和其他定制电路。它使用运动检测算法连续监测图像帧,并在白天或晚上检测到移动的动物时开始记录。田间试验表明,该系统成功记录了大量动物。使用人工生成的运动进行的实验室测试表明,该系统以0.99的高精度成功记录了视频帧,并提供了5.208W的优化峰值功耗。在现场试验中,共有27台摄像机节省了85,870个视频片段,其中423个视频片段成功记录了放热动物(爬行动物,两栖动物,和节肢动物)。这种新开发的相机陷阱将使野生动物生物学家受益,因为它成功地监测了吸热和放热的动物。
    Current camera traps use passive infrared triggers; therefore, they only capture images when animals have a substantially different surface body temperature than the background. Endothermic animals, such as mammals and birds, provide adequate temperature contrast to trigger cameras, while ectothermic animals, such as amphibians, reptiles, and invertebrates, do not. Therefore, a camera trap that is capable of monitoring ectotherms can expand the capacity of ecological research on ectothermic animals. This study presents the design, development, and evaluation of a solar-powered and artificial-intelligence-assisted camera trap system with the ability to monitor both endothermic and ectothermic animals. The system is developed using a central processing unit, integrated graphics processing unit, camera, infrared light, flash drive, printed circuit board, solar panel, battery, microphone, GPS receiver, temperature/humidity sensor, light sensor, and other customized circuitry. It continuously monitors image frames using a motion detection algorithm and commences recording when a moving animal is detected during the day or night. Field trials demonstrate that this system successfully recorded a high number of animals. Lab testing using artificially generated motion demonstrated that the system successfully recorded within video frames at a high accuracy of 0.99, providing an optimized peak power consumption of 5.208 W. No water or dust entered the cases during field trials. A total of 27 cameras saved 85,870 video segments during field trials, of which 423 video segments successfully recorded ectothermic animals (reptiles, amphibians, and arthropods). This newly developed camera trap will benefit wildlife biologists, as it successfully monitors both endothermic and ectothermic animals.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    COVID-19限制了户外运动和住院接受康复治疗。以家庭为基础的培训和康复教练系统已经成为克服这些情况的一种方法。传统的光学运动捕捉系统,比如VICON,已用于测量精确的运动,并在锻炼或康复过程中提供姿势反馈;然而,由于其成本和空间要求高,因此其应用仅限于专业设施。为了将适用性扩展到基于家庭的使用,我们设计了可穿戴皮肤标记(WSM),具有可以通过低成本网络摄像机检测到的特定于身体部位的模式。WSM是带状且可拉伸的,因此可以像布料一样佩戴,以最小的努力放置。特定于身体段的模式实现了实时数据处理,这减少了标记数据的后处理时间。通过识别段特定的图案来获得每个WSM的6自由度(DOF)姿势;然后利用通过三角测量找到的图案的轮廓角的3D配置来构造每个WSM的坐标。通过三个实验对WSM系统进行了验证。通过测量WSM的假阳性和假阴性率来评估标记识别的鲁棒性。对于准确性验证,获得了3自由度万向节的机械关节和行走的人类受试者的下肢关节的角度估计结果,并与参考系统进行了比较。包括万向节实验,以评估我们的系统在没有皮肤运动伪影的情况下的准确性。WSM和编码器之间的差异的最大标准偏差为0.9°的万向节实验,对于人体实验,WSM和VICON之间的角度为5.0°。精度与参考系统相当,使其适合家庭环境应用。
    COVID-19 has restricted outdoor exercise and hospital visits for rehabilitation therapy. Home-based training and rehabilitation coaching systems have emerged as a way to overcome these circumstances. Conventional optical motion-capture systems, such as VICON, have been used for measuring precise movement and providing posture feedback during exercise or rehabilitation; however, its application is limited to professional facilities because of its high cost and space requirement. To extend the applicability to home-based use, we designed wearable skin markers (WSMs) with body segment-specific patterns that can be detected by low-cost web cameras. WSMs are band-shaped and stretchable and thus can be worn like cloth, with minimal effort for placement. The body segment-specific patterns enable real-time data processing, which reduces the marker data post-processing time. A 6-degree-of-freedom (DOF) pose for each WSM is obtained by recognizing the segment-specific patterns; the 3D configuration of the contoured corners of the patterns found by triangulation is then utilized to construct the coordinates of each WSM. The WSM system was validated via three experiments. The robustness of marker recognition was evaluated by measuring the false-positive and false-negative rates of WSM. For accuracy validation, the angle estimation results were obtained for the mechanical joint of a 3-DOF gimbal and lower-limb joints of a walking human subject and compared to the reference systems. The gimbal experiment was included to evaluate the accuracy of our system in the condition with no skin movement artifact. The maximum standard deviation of the difference between WSM and the encoder was 0.9 °   for the gimbal experiment, and that between WSM and VICON was 5.0 ° for the human experiment. The accuracy was comparable to the reference systems, making it suitable for home environment application.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    浮游生物数量的变化会影响整个海洋生态系统。准确评估浮游生物的动态演变对监测海洋环境和全球气候变化具有重要意义。在本文中,介绍了一种利用水下机器人平台进行海洋生态系统深海浮游生物群落检测的新方法。这些视频是在距离海底1.5米的地方采样的,焦距为1.5-2.5m。光流场用于检测浮游生物群落。我们表明,对于在两个连续视频帧中在空间上不重叠的每个移动浮游生物,在两个连续的光流场中,浮游生物空间位置的时间梯度彼此相反。Further,横向和垂直梯度在两个连续的光流场中具有相同的值和方向。因此,在深海环境中复杂的动态背景下,可以准确地检测出移动浮游生物。与手动地面实况的实验比较充分验证了所提出方法的有效性,优于六种最先进的方法。
    Variations in the quantity of plankton impact the entire marine ecosystem. It is of great significance to accurately assess the dynamic evolution of the plankton for monitoring the marine environment and global climate change. In this paper, a novel method is introduced for deep-sea plankton community detection in marine ecosystem using an underwater robotic platform. The videos were sampled at a distance of 1.5 m from the ocean floor, with a focal length of 1.5-2.5 m. The optical flow field is used to detect plankton community. We showed that for each of the moving plankton that do not overlap in space in two consecutive video frames, the time gradient of the spatial position of the plankton are opposite to each other in two consecutive optical flow fields. Further, the lateral and vertical gradients have the same value and orientation in two consecutive optical flow fields. Accordingly, moving plankton can be accurately detected under the complex dynamic background in the deep-sea environment. Experimental comparison with manual ground-truth fully validated the efficacy of the proposed methodology, which outperforms six state-of-the-art approaches.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    进行了可行性研究,以研究在半结构化室内环境中使用低成本可穿戴相机对步态速度进行分类的可穿戴步态分析系统的使用。数据是从19名参与者那里收集的,这些参与者在室内步行序列中以不同的自定速度(慢速,中等,和快速)。使用该系统的步态参数与从包括单个三轴加速度计的背心和从基于标记的光学运动捕获系统获得的参数进行比较。使用计算机视觉技术和信号处理方法从每个步态记录设备生成频域步态参数,并对这些参数进行了分析,以确定不同测量系统在区分步态速度方面的有效性。结果表明,作者的低成本,便携式,基于视觉的系统可以有效地用于家庭步态分析。
    A feasibility study was conducted to investigate the use of a wearable gait analysis system for classifying gait speed using a low-cost wearable camera in a semi-structured indoor setting. Data were collected from 19 participants who wore the system during indoor walk sequences at varying self-determined speeds (slow, medium, and fast). Gait parameters using this system were compared with parameters obtained from a vest comprising of a single triaxial accelerometer and from a marker-based optical motion-capture system. Computer-vision techniques and signal processing methods were used to generate frequency-domain gait parameters from each gait-recording device, and those parameters were analysed to determine the effectiveness of the different measurement systems in discriminating gait speed. Results indicate that the authors\' low-cost, portable, vision-based system can be effectively used for in-home gait analysis.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Sci-hub)

       PDF(Pubmed)

  • 文章类型: Journal Article
    在腹腔镜检查视频中自动识别仪器提出了许多需要解决的挑战,比如识别以各种表示和不同照明条件出现的多种乐器,反过来可能会被其他仪器挡住,组织,血,或烟雾。考虑到这些挑战,例如,对于识别方法而言,首先在视频帧序列中检测仪器帧以用于进一步仅调查这些帧可能是有益的。此预识别步骤也与腹腔镜检查视频中的许多其他分类任务有关,如行动识别或不良事件分析。在这项工作中,作者解决了二进制分类的任务,以将视频帧识别为仪器或非仪器图像。他们研究卷积神经网络模型,以学习视频中仪器帧的表示,并仔细研究学习的激活模式。对于这项任务,使用公开可用的仪器计数分类数据集对GoogLeNet和批量标准化进行训练和验证。他们将迁移学习与从头开始学习进行了比较,并对胆囊切除术和妇科的数据集进行了评估。评估表明,与从头开始训练模型相比,在仪器和非仪器图像上微调预训练模型的学习速度快得多,也更稳定。
    Automatic recognition of instruments in laparoscopy videos poses many challenges that need to be addressed, like identifying multiple instruments appearing in various representations and in different lighting conditions, which in turn may be occluded by other instruments, tissue, blood, or smoke. Considering these challenges, it may be beneficial for recognition approaches that instrument frames are first detected in a sequence of video frames for further investigating only these frames. This pre-recognition step is also relevant for many other classification tasks in laparoscopy videos, such as action recognition or adverse event analysis. In this work, the authors address the task of binary classification to recognise video frames as either instrument or non-instrument images. They examine convolutional neural network models to learn the representation of instrument frames in videos and take a closer look at learned activation patterns. For this task, GoogLeNet together with batch normalisation is trained and validated using a publicly available dataset for instrument count classifications. They compared transfer learning with learning from scratch and evaluate on datasets from cholecystectomy and gynaecology. The evaluation shows that fine-tuning a pre-trained model on the instrument and non-instrument images is much faster and more stable in learning than training a model from scratch.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Sci-hub)

       PDF(Pubmed)

  • 文章类型: Journal Article
    Depth estimation plays an important role in vision-based laparoscope surgical navigation systems. Most learning-based depth estimation methods require ground truth depth or disparity images for training; however, these data are difficult to obtain in laparoscopy. The authors present an unsupervised learning depth estimation approach by fusing traditional stereo knowledge. The traditional stereo method is used to generate proxy disparity labels, in which unreliable depth measurements are removed via a confidence measure to improve stereo accuracy. The disparity images are generated by training a dual encoder-decoder convolutional neural network from rectified stereo images coupled with proxy labels generated by the traditional stereo method. A principled mask is computed to exclude the pixels, which are not seen in one of views due to parallax effects from the calculation of loss function. Moreover, the neighbourhood smoothness term is employed to constrain neighbouring pixels with similar appearances to generate a smooth depth surface. This approach can make the depth of the projected point cloud closer to the real surgical site and preserve realistic details. The authors demonstrate the performance of the method by training and evaluation with a partial nephrectomy da Vinci surgery dataset and heart phantom data from the Hamlyn Centre.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    重复图像特征跟踪(RIFT)通常用于从图像对测量冰川表面运动,最常利用归一化互相关(NCC)。多图像多芯片(MIMC)算法成功地采用冗余匹配(即,使用不同的设置组合在每个区域上重复匹配过程)以提高匹配成功率。由于大量的重复计算,然而,原始的MIMC算法速度较慢,在高剪切流区域仍然容易失效。在这里,我们介绍了MIMC算法的几个主要更新,以提高速度和匹配成功率。首先,我们通过交换图像顺序和匹配方向来包括额外的冗余测量;我们称之为Quadramatching的过程。第二,我们利用先验的冰速度信息通过我们称为动态线性约束(DLC)的系统来限制NCC搜索空间,这大大减少了计算时间并提高了成功匹配的速率。此外,我们开发了一种新颖的后处理算法,伪平滑,以确定最可能的位移。我们的测试揭示了这些升级在整体MIMC性能方面的互补性和多重性。
    Repeat Image Feature Tracking (RIFT) is commonly used to measure glacier surface motion from pairs of images, most often utilizing normalized cross correlation (NCC). The Multiple-Image Multiple-Chip (MIMC) algorithm successfully employed redundant matching (i.e. repeating the matching process over each area using varying combinations of settings) to increase the matching success rate. Due to the large number of repeat calculations, however, the original MIMC algorithm was slow and still prone to failure in areas of high shearing flow. Here we present several major updates to the MIMC algorithm that increase both speed and matching success rate. First, we include additional redundant measurements by swapping the image order and matching direction; a process we term Quadramatching. Second, we utilize a priori ice velocity information to confine the NCC search space through a system we term dynamic linear constraint (DLC), which substantially reduces the computation time and increases the rate of successful matches. Additionally, we develop a novel post-processing algorithm, pseudosmoothing, to determine the most probable displacement. Our tests reveal the complimentary and multiplicative nature of these upgrades in their improvement in overall MIMC performance.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    Cerebrovascular surgery treats vessel abnormalities in the brain and spinal cord, including arteriovenous malformations (AVMs) and aneurysms. These procedures often involve clipping the vessels feeding blood to these abnormalities, making accurate classification of blood vessel types (feeding versus draining) important during surgery. Previous work to guide the intraoperative identification of the vessels included augmented reality (AR) using pre-operative images, injected dyes, and Doppler ultrasound, but each with their drawbacks. The authors propose and demonstrate a novel technique to help differentiate vessels by enhancing short videos of a few seconds from the surgical microscope using motion magnification and spectral analysis, and constructing AR views that fuse the analysis results as intuitive colourmaps and the surgical microscopic view. They demonstrated the proposed technique retrospectively with two real cerebrovascular surgical cases: one AVM and one aneurysm. The results showed that the proposed technique can help characterise different vessel types (feeding and draining the abnormality), which agree with those identified by the operating surgeon.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Sci-hub)

       PDF(Pubmed)

  • 文章类型: Journal Article
    OBJECTIVE: To examine the impact of 4D-PET on target volume delineation of upper-abdominal tumors, versus conventional un-gated PET.
    METHODS: Four patients with upper-abdominal tumors underwent respiratory-correlated FDG PET/CT scanning (4D-PET) as part of a continuing IRB-approved research protocol. Internal target volumes of FDG-avid tumors were contoured on the 4D-PET and conventional un-gated PET by a radiation oncologist who is a specialist in gastro-intestinal tumors. To create the 4D-PET ITV, the end-inhale and end-exhale 4D-PET phases were used. The relative volumes and volumetric overlaps of the 4D and un-gated target volumes were examined. Additionally, 4D-PET was used to measure the motion of the tumors.
    RESULTS: Of the four patients who were imaged, one showed minimal motion (〈 3 mm in any direction) and one showed minimal FDG avidity; these were removed from further analysis. Of the two tumors which showed significant motion and FDG uptake, 4D-PET volumes were 28% and 21% larger than un-gated PET volumes. The un-gated PET volumes were almost entirely contained within the 4D-PET volumes (95% and 93% for the two tumors). Tumors appeared to deform as well as translate with breathing, although this could be due to varying intra-gate motion rather than actual physiological deformation. The superior-inferior borders of the tumors exhibited the most motion, with displacements of 5.6 mm and 6.4 mm.
    CONCLUSIONS: 4D-PET can be used to estimate the motion of FDG-avid upper-abdominal tumors. Use of 4D-PET increases the size of target volumes compared to un-gated PET in a subset of upper-abdominal cancer patients. Direct measurement of tumor motion and deformation by 4D-PET imaging could allow the use of patient-specific margins rather than population-based margins, potentially leading to increased target coverage and reduced normal tissue irradiation.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

公众号