synthetic dataset

  • 文章类型: Journal Article
    个性化的饮食建议越来越受欢迎,然而,目前大多数方法都是基于个体的遗传和表型特征,而在很大程度上忽略了其他决定因素,如社会经济和认知变量。本文通过测试同时针对个人的社会人口群体量身定制的个性化健康饮食建议的有效性,提供了新颖的见解。认知特征,和感官偏好。我们首先使用现有数据基于来自3654个家庭的信息构建合成数据集(研究1a),然后开发了一个聚类模型来识别具有相似社会人口统计特征的个体,认知,和感官方面(研究1b)。最后,在研究2中,我们使用8个集群的特征来构建8个独立的个性化食物选择建议,并评估它们促进水果和蔬菜消费增加以及饱和脂肪和糖摄入量减少的能力。我们向218位参与者提供了通用的英国政府“EatWell”建议,为他们分配的集群量身定制的建议(匹配的个性化),或为不同的集群量身定制的建议(无与伦比的个性化)。结果表明,与一般建议相比,接受匹配的个性化建议的参与者更有可能表明他们会改变饮食.当参与者收到无与伦比的个性化建议时,他们同样有动机增加蔬菜消费量和减少饱和脂肪摄入量。可能凸显提供替代食物选择的力量。总的来说,这项研究表明,个性化食物选择建议的力量,基于个体特征的组合,在推动饮食改变方面比目前的方法更有效。我们的研究还强调了通过自动提供基于网络的个性化建议来解决人口健康问题的可行性。
    Personalised dietary advice has become increasingly popular, currently however most approaches are based on an individual\'s genetic and phenotypic profile whilst largely ignoring other determinants such as socio economic and cognitive variables. This paper provides novel insights by testing the effectiveness of personalised healthy eating advice concurrently tailored to an individual\'s socio-demographic group, cognitive characteristics, and sensory preferences. We first used existing data to build a synthetic dataset based on information from 3654 households (Study 1a), and then developed a cluster model to identify individuals characterised by similar socio-demographic, cognitive, and sensory aspects (Study 1b). Finally, in Study 2 we used the characteristics of 8 clusters to build 8 separate personalised food choice advice and assess their ability to motivate the increased consumption of fruit and vegetables and decreased intakes of saturated fat and sugar. We presented 218 participants with either generic UK Government \"EatWell\" advice, advice that was tailored to their allocated cluster (matched personalised), or advice tailored to a different cluster (unmatched personalised). Results showed that, when compared to generic advice, participants that received matched personalised advice were significantly more likely to indicate they would change their diet. Participants were similarly motivated to increase vegetable consumption and decrease saturated fat intake when they received unmatched personalised advice, potentially highlighting the power of providing alternative food choices. Overall, this study demonstrated that the power of personalizing food choice advice, based on a combination of individual characteristics, can be more effective than current approaches in motivating dietary change. Our study also emphasizes the viability of addressing population health through automatically delivered web-based personalised advice.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    通过基准标记对头部姿势的光学跟踪已经被证明能够在磁共振成像期间有效地校正大脑中的运动伪影,但是由于冗长的校准和设置时间而仍然难以在临床中实现。由于运动校正所需的亚毫米空间分辨率,用于无标记头部姿势估计的深度学习的进展尚未应用于该问题。在目前的工作中,描述了两个用于开发和训练神经网络的光学跟踪系统:一个基于标记的系统(用于测量地面真相头部姿势的测试平台)具有高跟踪保真度作为训练标签,和一个无标记的基于深度学习的系统,使用无标记的头部的图像作为网络的输入。无标记系统有可能克服标记物遮挡的问题,标记的刚性连接不足,冗长的校准时间,以及跨自由度(DOF)的不平等性能,所有这些都阻碍了在临床中采用基于标记的解决方案。提供了有关用作地面实况的自定义莫尔增强基准标记的开发以及两个光学跟踪系统的校准程序的详细信息。此外,描述了合成头部姿态数据集的开发,以证明简单卷积神经网络的概念和初始预训练。结果表明,地面实况系统已得到充分校准,可以跟踪磁头姿态,误差<1mm和<1°。跟踪健康的数据,成人参与者显示。预训练结果表明,在训练数据集包含和排除的头部模型上,6个自由度的平均均方根误差为0.13和0.36(mm或度)。分别。总的来说,这项工作表明了基于深度学习的方法的出色可行性,并将使未来的工作能够在MRI环境中对真实数据集进行培训和测试。
    Optical tracking of head pose via fiducial markers has been proven to enable effective correction of motion artifacts in the brain during magnetic resonance imaging but remains difficult to implement in the clinic due to lengthy calibration and set up times. Advances in deep learning for markerless head pose estimation have yet to be applied to this problem because of the sub-millimetre spatial resolution required for motion correction. In the present work, two optical tracking systems are described for the development and training of a neural network: one marker-based system (a testing platform for measuring ground truth head pose) with high tracking fidelity to act as the training labels, and one markerless deep-learning-based system using images of the markerless head as input to the network. The markerless system has the potential to overcome issues of marker occlusion, insufficient rigid attachment of the marker, lengthy calibration times, and unequal performance across degrees of freedom (DOF), all of which hamper the adoption of marker-based solutions in the clinic. Detail is provided on the development of a custom moiré-enhanced fiducial marker for use as ground truth and on the calibration procedure for both optical tracking systems. Additionally, the development of a synthetic head pose dataset is described for the proof of concept and initial pre-training of a simple convolutional neural network. Results indicate that the ground truth system has been sufficiently calibrated and can track head pose with an error of <1 mm and <1°. Tracking data of a healthy, adult participant are shown. Pre-training results show that the average root-mean-squared error across the 6 DOF is 0.13 and 0.36 (mm or degrees) on a head model included and excluded from the training dataset, respectively. Overall, this work indicates excellent feasibility of the deep-learning-based approach and will enable future work in training and testing on a real dataset in the MRI environment.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    疾病预测受到与真实医疗数据相关的数据集的稀缺性和隐私问题的极大挑战。克服这一障碍的一种方法是使用使用生成对抗网络(GAN)生成的合成数据。GAN可以增加数据量,同时生成与个人信息没有直接链接的合成数据集。本研究率先使用GAN创建合成数据集和使用传统增强技术增强的数据集来完成我们的二元分类任务。这项研究的主要目的是评估我们的新条件深度卷积神经网络(C-DCNN)模型在通过利用这些增强和合成数据集对脑肿瘤进行分类方面的性能。我们利用了先进的GAN模型,包括条件深度卷积生成对抗网络(DCGAN),生成合成数据,保留原始数据集的基本特征,同时确保隐私保护。我们的C-DCNN模型在增强和合成数据集上进行了训练,并且其性能以最先进的型号为基准,例如ResNet50、VGG16、VGG19和InceptionV3。评估指标表明,我们的C-DCNN模型达到了准确性,精度,召回,F1在合成和增强图像上的得分为99%,优于比较模型。这项研究的结果强调了使用GAN生成的合成数据来增强用于医学图像分类的机器学习模型的训练的潜力。特别是在可用数据有限的情况下。这种方法不仅提高了模型的准确性,而且解决了隐私问题,使其成为疾病预测和诊断的现实世界临床应用的可行解决方案。
    Disease prediction is greatly challenged by the scarcity of datasets and privacy concerns associated with real medical data. An approach that stands out to circumvent this hurdle is the use of synthetic data generated using Generative Adversarial Networks (GANs). GANs can increase data volume while generating synthetic datasets that have no direct link to personal information. This study pioneers the use of GANs to create synthetic datasets and datasets augmented using traditional augmentation techniques for our binary classification task. The primary aim of this research was to evaluate the performance of our novel Conditional Deep Convolutional Neural Network (C-DCNN) model in classifying brain tumors by leveraging these augmented and synthetic datasets. We utilized advanced GAN models, including Conditional Deep Convolutional Generative Adversarial Network (DCGAN), to produce synthetic data that retained essential characteristics of the original datasets while ensuring privacy protection. Our C-DCNN model was trained on both augmented and synthetic datasets, and its performance was benchmarked against state-of-the-art models such as ResNet50, VGG16, VGG19, and InceptionV3. The evaluation metrics demonstrated that our C-DCNN model achieved accuracy, precision, recall, and F1 scores of 99% on both synthetic and augmented images, outperforming the comparative models. The findings of this study highlight the potential of using GAN-generated synthetic data in enhancing the training of machine learning models for medical image classification, particularly in scenarios with limited data available. This approach not only improves model accuracy but also addresses privacy concerns, making it a viable solution for real-world clinical applications in disease prediction and diagnosis.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    影子,由于没有光导致的自然现象,在农业中起着举足轻重的作用,特别是在植物光合作用等过程中。尽管有通用影子数据集,许多人遭受注释错误,并且缺乏内部可能存在人类活动的农业阴影的详细表示,不包括来自卫星或无人机视图的那些。在本文中,我们提供了一个综合生成的自上而下的阴影分割数据集的评估,其特征是逼真的渲染和精确的阴影掩模。我们的目标是确定其与现实世界数据集相比的功效,并评估注释质量和图像域等因素如何影响神经网络模型训练。要建立基线,我们训练了许多基线架构,随后使用各种免费的影子数据集探索了迁移学习。与其他阴影数据集的训练集相比,我们进一步评估了域外性能。我们的研究结果表明,AgroSegNet表现出竞争力,对迁移学习是有效的,特别是在类似于农业的领域。
    Shadow, a natural phenomenon resulting from the absence of light, plays a pivotal role in agriculture, particularly in processes such as photosynthesis in plants. Despite the availability of generic shadow datasets, many suffer from annotation errors and lack detailed representations of agricultural shadows with possible human activity inside, excluding those derived from satellite or drone views. In this paper, we present an evaluation of a synthetically generated top-down shadow segmentation dataset characterized by photorealistic rendering and accurate shadow masks. We aim to determine its efficacy compared to real-world datasets and assess how factors such as annotation quality and image domain influence neural network model training. To establish a baseline, we trained numerous baseline architectures and subsequently explored transfer learning using various freely available shadow datasets. We further evaluated the out-of-domain performance compared to the training set of other shadow datasets. Our findings suggest that AgroSegNet demonstrates competitive performance and is effective for transfer learning, particularly in domains similar to agriculture.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    计算机视觉中的一些研究已经检查了镜面去除,这对于目标检测和识别至关重要。这项研究传统上分为两个任务:镜面高光去除,重点是去除物体表面的镜面高光,和反射去除,处理发生在玻璃表面的镜面反射。在现实中,然而,这两种类型的镜面效应经常共存,使其成为尚未得到充分解决的根本挑战。认识到集成在这两个任务中处理的镜面组件的必要性,我们构建了一个镜面光(S-Light)DB,用于训练基于单图像的深度学习模型。此外,考虑到缺乏定量评估的基准数据集,多尺度归一化互相关(MS-NCC)度量,它考虑了镜面和漫反射分量之间的相关性,被引入来评估学习成果。
    Several studies in computer vision have examined specular removal, which is crucial for object detection and recognition. This research has traditionally been divided into two tasks: specular highlight removal, which focuses on removing specular highlights on object surfaces, and reflection removal, which deals with specular reflections occurring on glass surfaces. In reality, however, both types of specular effects often coexist, making it a fundamental challenge that has not been adequately addressed. Recognizing the necessity of integrating specular components handled in both tasks, we constructed a specular-light (S-Light) DB for training single-image-based deep learning models. Moreover, considering the absence of benchmark datasets for quantitative evaluation, the multi-scale normalized cross correlation (MS-NCC) metric, which considers the correlation between specular and diffuse components, was introduced to assess the learning outcomes.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    车辆检测是计算机视觉应用于航空和卫星图像的一个非常重要的方面,促进实例计数等活动,速度估计,交通预测,等。准确车辆检测的可行性通常取决于有限的训练数据集,在收集和注释任务中需要大量的手动工作。此外,没有已知的公开可用数据集。我们的目标是在Blender软件中从航拍图像和3D模型构建合成数据集生成管道。数据集生成流水线由七个步骤组成,并且产生具有YOLO和coco格式的边界框的期望数量的图像。该合成数据集是按照该流水线中描述的步骤产生的。它由5000个2048×2048图像组成,在没有来自世界各地的汽车的图像中,汽车插入道路和高速公路。我们认为,这个数据集和相应的管道可能对车辆检测非常重要,促进模型对特定需求和上下文的可定制性。
    Vehicle detection is a very important aspect of computer vision application to aerial and satellite imagery, facilitating activities such as instance counting, velocity estimation, traffic predictions, etc. The feasibility of accurate vehicle detection often depends on limited training datasets, requiring a lot of manual work in collection and annotation tasks. Furthermore, there are no known publicly available datasets. Our aim was to construct a pipeline for synthetic dataset generation from aerial imagery and 3D models in Blender software. The dataset generation pipeline consists of seven steps and results in a wished number of images with bounding boxes in YOLO and coco formats. This synthetic dataset has been produced following the steps described in this pipeline. It consists of 5000 2048 × 2048 images with cars inserted into the roads and highways at the images without cars from all over the world. We believe that this dataset and the respective pipeline might be of great importance for vehicle detection, facilitating the customizability of the models to specific needs and context.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    视觉定位是指通过分析查询图像和预先存在的图像集之间的空间关系来确定观察者的姿势的过程。在这个过程中,识别图像之间的匹配视觉特征并将其用于姿态估计;因此,估计的准确性在很大程度上依赖于特征匹配的精度。不正确的功能匹配,例如图像中不同对象和/或对象内不同点之间的那些,因此应该避免。在本文中,我们的初始评估集中在测量图像数据集中每个对象类的可靠性有关的姿态估计精度。这项评估表明,建筑类是可靠的,而人类在不同地点表现出不可靠性。随后的研究通过人为增加不可靠物体-人类的比例,更深入地研究了姿势估计精度的下降。研究结果表明,当图像中人类的平均比例超过20%时,出现了显着的下降。我们讨论了用于视觉定位的数据集构建的结果和意义。
    Visual localization refers to the process of determining an observer\'s pose by analyzing the spatial relationships between a query image and a pre-existing set of images. In this procedure, matched visual features between images are identified and utilized for pose estimation; consequently, the accuracy of the estimation heavily relies on the precision of feature matching. Incorrect feature matchings, such as those between different objects and/or different points within an object in an image, should thus be avoided. In this paper, our initial evaluation focused on gauging the reliability of each object class within image datasets concerning pose estimation accuracy. This assessment revealed the building class to be reliable, while humans exhibited unreliability across diverse locations. The subsequent study delved deeper into the degradation of pose estimation accuracy by artificially increasing the proportion of the unreliable object-humans. The findings revealed a noteworthy decline started when the average proportion of the humans in the images exceeded 20%. We discuss the results and implications for dataset construction for visual localization.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    本研究演示了如何通过少量图像生成三维(3D)身体模型,并使用生成的3D身体数据得出与实际值相似的身体值。在这项研究中,使用用手机拍摄的正面和侧面的两张全身图片开发了可用于身体类型诊断的3D身体模型。对于数据训练,使用了SizeKorea提供的400个3D身体数据集(男性:200,女性:200),和四个模型,即,三维递归重建神经网络,点云生成对抗网络,皮纳多人线性模型,和像素对齐冲击功能,用于高分辨率3D人体数字化,被使用。本研究提出的模型进行了分析和比较。共分析了10名男女,通过将从2D图像输入得出的3D身体数据与使用身体扫描仪获得的数据进行比较,验证了它们相应的3D模型。通过从2D图像导出的3D数据与使用实际身体扫描仪导出的3D数据之间的差异来验证模型。与本研究中无法用于得出身体值的3D生成模型不同,所提出的模型被成功地用于推导各种身体值,表明该模型可以用于识别各种身体类型并监测未来的肥胖。
    This study demonstrates how to generate a three-dimensional (3D) body model through a small number of images and derive body values similar to the actual values using generated 3D body data. In this study, a 3D body model that can be used for body type diagnosis was developed using two full-body pictures of the front and side taken with a mobile phone. For data training, 400 3D body datasets (male: 200, female: 200) provided by Size Korea were used, and four models, i.e., 3D recurrent reconstruction neural network, point cloud generative adversarial network, skinned multi-person linear model, and pixel-aligned impact function for high-resolution 3D human digitization, were used. The models proposed in this study were analyzed and compared. A total of 10 men and women were analyzed, and their corresponding 3D models were verified by comparing 3D body data derived from 2D image inputs with those obtained using a body scanner. The model was verified through the difference between 3D data derived from the 2D image and those derived using an actual body scanner. Unlike the 3D generation models that could not be used to derive the body values in this study, the proposed model was successfully used to derive various body values, indicating that this model can be implemented to identify various body types and monitor obesity in the future.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    由于难以生成6自由度(6-DoF)对象姿态估计数据集,合成数据和真实数据之间存在域间隙,现有的姿态估计方法在提高准确性和泛化方面面临挑战。本文提出了一种方法,采用更高质量的数据集和基于深度学习的方法来减少合成数据和真实数据之间的领域差距问题,并提高姿态估计的准确性。从Blenderproc获得高质量数据集,并使用双边滤波对其进行创新性处理以减少差距。提出了一种新的基于注意力掩模区域的卷积神经网络(R-CNN),以降低计算成本并提高模型检测精度。同时,通过添加一层自下而上的路径来提取底层特征的内化,从而实现了改进的特征金字塔网络(iFPN)。因此,提出了一种新的卷积块注意模块-卷积去噪自动编码器(CBAM-CDAE)网络,通过呈现通道注意和空间注意机制来提高AE提取图像特征的能力。最后,通过姿态细化获得准确的6-DoF物体姿态。将所提出的方法与使用T-LESS和LineMOD数据集的其他模型进行比较。比较结果表明,该方法优于其他估计模型。
    Due to the difficulty in generating a 6-Degree-of-Freedom (6-DoF) object pose estimation dataset, and the existence of domain gaps between synthetic and real data, existing pose estimation methods face challenges in improving accuracy and generalization. This paper proposes a methodology that employs higher quality datasets and deep learning-based methods to reduce the problem of domain gaps between synthetic and real data and enhance the accuracy of pose estimation. The high-quality dataset is obtained from Blenderproc and it is innovatively processed using bilateral filtering to reduce the gap. A novel attention-based mask region-based convolutional neural network (R-CNN) is proposed to reduce the computation cost and improve the model detection accuracy. Meanwhile, an improved feature pyramidal network (iFPN) is achieved by adding a layer of bottom-up paths to extract the internalization of features of the underlying layer. Consequently, a novel convolutional block attention module-convolutional denoising autoencoder (CBAM-CDAE) network is proposed by presenting channel attention and spatial attention mechanisms to improve the ability of AE to extract images\' features. Finally, an accurate 6-DoF object pose is obtained through pose refinement. The proposed approach is compared to other models using the T-LESS and LineMOD datasets. Comparison results demonstrate the proposed approach outperforms the other estimation models.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    C.elegans在图像序列中的姿态估计是具有挑战性的,并且在低分辨率图像中甚至更加困难。问题包括遮挡,蠕虫身份的丧失,与过于复杂或难以解决的聚合重叠,即使是人眼。神经网络,另一方面,在低分辨率和高分辨率图像中都显示出良好的效果。然而,神经网络模型中的训练需要非常大且平衡的数据集,这有时是不可能的或太贵了。在这篇文章中,提出了一种在多蠕虫聚集和噪声聚集情况下预测秀丽隐杆线虫姿态的新方法。为了解决这个问题,我们使用了一种改进的U-Net模型,该模型能够获得下一个聚合蠕虫姿势的图像。使用具有合成图像模拟器的定制生成的数据集来训练/验证该神经网络模型。随后,用真实图像的数据集进行测试。获得的结果在精度上大于75%,在交集(IoU)值的情况下大于0.65。
    Pose estimation of C. elegans in image sequences is challenging and even more difficult in low-resolution images. Problems range from occlusions, loss of worm identity, and overlaps to aggregations that are too complex or difficult to resolve, even for the human eye. Neural networks, on the other hand, have shown good results in both low-resolution and high-resolution images. However, training in a neural network model requires a very large and balanced dataset, which is sometimes impossible or too expensive to obtain. In this article, a novel method for predicting C. elegans poses in cases of multi-worm aggregation and aggregation with noise is proposed. To solve this problem we use an improved U-Net model capable of obtaining images of the next aggregated worm posture. This neural network model was trained/validated using a custom-generated dataset with a synthetic image simulator. Subsequently, tested with a dataset of real images. The results obtained were greater than 75% in precision and 0.65 with Intersection over Union (IoU) values.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

公众号