data processing

数据处理
  • 文章类型: Journal Article
    本文提出的研究的主要目的是介绍一种比传统深度神经网络需要更少的计算能力的人工神经网络。该神经网络的开发是通过应用有序模糊数(OFN)实现的。在工业4.0的背景下,有许多应用可以利用该解决方案进行数据处理。它允许在小型设备上的网络边缘部署人工智能,无需将大量数据传输到云服务器进行分析。这样的网络将更容易在小规模解决方案中实现,比如物联网,在未来。本文介绍了对真实系统进行监控的测试结果,异常被检测和预测。
    The primary objective of the research presented in this article is to introduce an artificial neural network that demands less computational power than a conventional deep neural network. The development of this ANN was achieved through the application of Ordered Fuzzy Numbers (OFNs). In the context of Industry 4.0, there are numerous applications where this solution could be utilized for data processing. It allows the deployment of Artificial Intelligence at the network edge on small devices, eliminating the need to transfer large amounts of data to a cloud server for analysis. Such networks will be easier to implement in small-scale solutions, like those for the Internet of Things, in the future. This paper presents test results where a real system was monitored, and anomalies were detected and predicted.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    作为数字表型,从智能手机等消费设备捕获主动和被动数据,变得更加普遍,正确处理数据并从中获得可复制功能的需求变得至关重要。Cortex是用于数字表型数据的开源数据处理管道,针对mindLAMP应用程序的使用进行了优化,全世界近100个研究团队都在使用它。Cortex旨在帮助团队(1)实时评估数字表型数据质量,(2)从数据中得出可复制的临床特征,和(3)实现易于共享的数据可视化。Cortex提供了许多选项来处理数字表型数据,尽管一些常见的方法可能对所有使用它的团队都有价值。本文强调了推理,代码,以及以简化方式充分处理数字表型数据所需的示例步骤。涵盖如何处理数据,评估其质量,派生特征,可视化发现,本文旨在为读者提供适用于分析任何数字表型数据集的知识和技能。更具体地说,本文将向读者传授CortexPython包的来龙去脉。这包括其与mindLAMP平台互动的背景信息,一些基本的命令来学习什么数据可以提取,和更高级的使用软件包与基本的Python混合,目标是创建一个相关矩阵。教程之后,讨论了Cortex的不同用例,连同限制。为了突出临床应用,本文还提供了3种简单的方法来实现在现实世界中使用Cortex的例子。通过了解如何使用数字表型数据并使用Cortex提供可部署的代码,这篇论文旨在展示数字表型的新领域如何既可以被所有人访问,又可以被严格的方法论。
    As digital phenotyping, the capture of active and passive data from consumer devices such as smartphones, becomes more common, the need to properly process the data and derive replicable features from it has become paramount. Cortex is an open-source data processing pipeline for digital phenotyping data, optimized for use with the mindLAMP apps, which is used by nearly 100 research teams across the world. Cortex is designed to help teams (1) assess digital phenotyping data quality in real time, (2) derive replicable clinical features from the data, and (3) enable easy-to-share data visualizations. Cortex offers many options to work with digital phenotyping data, although some common approaches are likely of value to all teams using it. This paper highlights the reasoning, code, and example steps necessary to fully work with digital phenotyping data in a streamlined manner. Covering how to work with the data, assess its quality, derive features, and visualize findings, this paper is designed to offer the reader the knowledge and skills to apply toward analyzing any digital phenotyping data set. More specifically, the paper will teach the reader the ins and outs of the Cortex Python package. This includes background information on its interaction with the mindLAMP platform, some basic commands to learn what data can be pulled and how, and more advanced use of the package mixed with basic Python with the goal of creating a correlation matrix. After the tutorial, different use cases of Cortex are discussed, along with limitations. Toward highlighting clinical applications, this paper also provides 3 easy ways to implement examples of Cortex use in real-world settings. By understanding how to work with digital phenotyping data and providing ready-to-deploy code with Cortex, the paper aims to show how the new field of digital phenotyping can be both accessible to all and rigorous in methodology.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    配电网在发展过程中的长期损耗是由于配电网管理模式落后造成的。传统的配电网网损分析计算方法已不能适应当前配电网的发展环境。为了提高在电力负荷数据中填充缺失值的准确性,提出了粒子群优化算法来优化聚类算法的聚类中心。此外,原始孤立森林异常识别算法可用于检测负荷数据中的异常值,负荷数据的变异系数,提高了算法的识别精度。最后,本文介绍了一种基于广度优先的大数据背景下线损计算方法。以云南省玉溪市配电网系统为例,并进行了仿真实验。结果表明,在部分数据缺失的情况下,增强模糊C均值聚类算法的误差平均为-6.35,标准差为4.015。改进的孤立森林算法受试者在样本异常模糊情况下的特征曲线下面积为0.8586,以最小的下降,根据变异系数,通过分析的细化,发现馈线损失率为7.62%。结果表明,该技术可以快速准确地进行配电网线损分析,可以作为配电网线损管理的指导。
    The long-term loss of distribution network in the process of distribution network development is caused by the backward management mode of distribution network. The traditional analysis and calculation methods of distribution network loss can not adapt to the current development environment of distribution network. To improve the accuracy of filling missing values in power load data, particle swarm optimization algorithm is proposed to optimize the clustering center of the clustering algorithm. Furthermore, the original isolated forest anomaly recognition algorithm can be used to detect outliers in the load data, and the coefficient of variation of the load data is used to improve the recognition accuracy of the algorithm. Finally, this paper introduces a breadth-first-based method for calculating line loss in the context of big data. An example is provided using the distribution network system of Yuxi City in Yunnan Province, and a simulation experiment is carried out. And the findings revealed that the error of the enhanced fuzzy C-mean clustering algorithm was on average - 6.35, with a standard deviation of 4.015 in the situation of partially missing data. The area under the characteristic curve of the improved isolated forest algorithm subjects in the case of the abnormal sample fuzzy situation was 0.8586, with the smallest decrease, based on the coefficient of variation, and through the refinement of the analysis, it was discovered that the feeder line loss rate is 7.62%. It is confirmed that the suggested technique can carry out distribution network line loss analysis fast and accurately and can serve as a guide for managing distribution network line loss.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    低温电子显微镜的共同挑战,例如取向偏差,构象多样性,和3D错误分类,复杂的单粒子分析,并导致大量的资源支出。我们之前介绍了一种使用最大费雷特直径分布的计算机模拟方法,费雷特的签名,表征圆盘形样品的样品异质性。这里,我们扩展了Feret签名方法,以确定包含任意形状且仅需要约1000个颗粒的样品的首选方向。该方法使得能够实时调整数据采集参数,以用于优化数据收集策略或帮助决定中断无效成像会话。除了检测首选方向,Feret签名方法可以作为初始图像处理步骤中分类不一致的早期预警系统,一种允许在数据处理中进行战略调整的能力。这些特征将Feret签名确立为在单粒子分析的背景下的有价值的辅助工具。显著加快了结构确定过程。
    Common challenges in cryogenic electron microscopy, such as orientation bias, conformational diversity, and 3D misclassification, complicate single particle analysis and lead to significant resource expenditure. We previously introduced an in silico method using the maximum Feret diameter distribution, the Feret signature, to characterize sample heterogeneity of disc-shaped samples. Here, we expanded the Feret signature methodology to identify preferred orientations of samples containing arbitrary shapes with only about 1000 particles required. This method enables real-time adjustments of data acquisition parameters for optimizing data collection strategies or aiding in decisions to discontinue ineffective imaging sessions. Beyond detecting preferred orientations, the Feret signature approach can serve as an early-warning system for inconsistencies in classification during initial image processing steps, a capability that allows for strategic adjustments in data processing. These features establish the Feret signature as a valuable auxiliary tool in the context of single particle analysis, significantly accelerating the structure determination process.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    本研究旨在衡量在三种不同的数据处理协议(原始,自定义处理,制造商处理)。距离的估计,使用10HzGNSS跟踪技术设备从一支球队的38名澳大利亚精英足球运动员的14场比赛中收集速度和加速度。从各自的专有软件导出原始和制造商处理的数据,并且针对三种处理方法计算两个常见的汇总加速度度量(中/高强度区内的努力次数和距离)。为了估计三种不同的数据处理方法对汇总指标的影响,使用线性混合模型。主要发现表明,三种处理方法之间存在实质性差异;制造商处理的加速度数据具有最低的报告距离(低184倍)和努力(低89倍),其次是定制处理距离(低3.3倍)和努力(低4.3倍),原始数据报告的距离和努力程度最高。结果表明,不同的处理方法改变了度量输出,进而改变了运动需求的量化(体积,度量的强度和频率)。教练,从业者和研究人员需要了解各种处理方法会改变加速度数据的汇总指标。通过了解这些指标如何受到处理方法的影响,他们可以更好地解释可用的数据,并有效地定制他们的培训计划,以满足比赛的需求。
    This study aimed to measure the differences in commonly used summary acceleration metrics during elite Australian football games under three different data processing protocols (raw, custom-processed, manufacturer-processed). Estimates of distance, speed and acceleration were collected with a 10-Hz GNSS tracking technology device from fourteen matches of 38 elite Australian football players from one team. Raw and manufacturer-processed data were exported from respective proprietary software and two common summary acceleration metrics (number of efforts and distance within medium/high-intensity zone) were calculated for the three processing methods. To estimate the effect of the three different data processing methods on the summary metrics, linear mixed models were used. The main findings demonstrated that there were substantial differences between the three processing methods; the manufacturer-processed acceleration data had the lowest reported distance (up to 184 times lower) and efforts (up to 89 times lower), followed by the custom-processed distance (up to 3.3 times lower) and efforts (up to 4.3 times lower), where raw data had the highest reported distance and efforts. The results indicated that different processing methods changed the metric output and in turn alters the quantification of the demands of a sport (volume, intensity and frequency of the metrics). Coaches, practitioners and researchers need to understand that various processing methods alter the summary metrics of acceleration data. By being informed about how these metrics are affected by processing methods, they can better interpret the data available and effectively tailor their training programs to match the demands of competition.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    论文“使用胰岛素和碳水化合物的吸收模型和深度倾斜以提高血糖水平预测”(Sensors2021,21,5273)提出了一种新的方法来预测1型糖尿病(T1DM)患者的血糖水平。通过从原始碳水化合物和胰岛素数据建立指数模型来模拟体内的吸收,作者报道,在预测未来一小时的血糖水平时,模型的均方根误差(RMSE)从15.5mg/dL(原始)降低至9.2mg/dL(指数).在这篇评论中,我们证明了那篇论文中使用的实验技术是有缺陷的,使其结果和结论无效。具体来说,在审查了作者的代码之后,我们发现模型验证方案是错误的,即,来自相同时间间隔的训练和测试数据是混合的.这意味着参考论文中报告的RMSE数字没有准确衡量所提出方法的预测能力。我们通过适当隔离训练和测试数据来修复测量技术,我们发现他们的模型实际上比论文中报道的要糟糕得多。事实上,那篇论文中提出的模型似乎没有比预测未来血糖水平与当前水平相同的幼稚模型表现更好。
    The paper \"Using Absorption Models for Insulin and Carbohydrates and Deep Leaning to Improve Glucose Level Predictions\" (Sensors2021, 21, 5273) proposes a novel approach to predicting blood glucose levels for people with type 1 diabetes mellitus (T1DM). By building exponential models from raw carbohydrate and insulin data to simulate the absorption in the body, the authors reported a reduction in their model\'s root-mean-square error (RMSE) from 15.5 mg/dL (raw) to 9.2 mg/dL (exponential) when predicting blood glucose levels one hour into the future. In this comment, we demonstrate that the experimental techniques used in that paper are flawed, which invalidates its results and conclusions. Specifically, after reviewing the authors\' code, we found that the model validation scheme was malformed, namely, the training and test data from the same time intervals were mixed. This means that the reported RMSE numbers in the referenced paper did not accurately measure the predictive capabilities of the approaches that were presented. We repaired the measurement technique by appropriately isolating the training and test data, and we discovered that their models actually performed dramatically worse than was reported in the paper. In fact, the models presented in the that paper do not appear to perform any better than a naive model that predicts future glucose levels to be the same as the current ones.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    当前的工作提出了由四个扭曲的胶合板条组成的轻质梁元素的综合空间数据集的生成,通过在受控的实验室条件下应用运动结构(SfM)-多视图立体(MVS)摄影测量技术来实现。数据收集过程精心进行,以确保准确性和准确性,采用不同长度的比例尺。然后使用摄影测量软件对捕获的图像进行处理,导致点云的产生,网格,和纹理文件。这些数据文件表示不同网格尺寸(原始,高聚,中聚,和低聚),为3D可视化添加高级别的细节。该数据集具有巨大的重用潜力,并为数值建模的进一步研究提供了必要的资源,复杂结构的模拟,和训练机器学习算法。这些数据还可以作为新兴摄影测量方法和形式发现技术的验证集,尤其是涉及大变形和几何非线性的,特别是在结构工程领域。
    The current work presents the generation of a comprehensive spatial dataset of a lightweight beam element composed of four twisted plywood strips, achieved through the application of Structure-from-Motion (SfM) - Multi-view Stereo (MVS) photogrammetry techniques in controlled laboratory conditions. The data collection process was meticulously conducted to ensure accuracy and precision, employing scale bars of varying lengths. The captured images were then processed using photogrammetric software, leading to the creation of point clouds, meshes, and texture files. These data files represent the 3D model of the beam at different mesh sizes (raw, high-poly, medium-poly, and low-poly), adding a high level of detail to the 3D visualization. The dataset holds significant reuse potential and offers essential resources for further studies in numerical modeling, simulations of complex structures, and training machine learning algorithms. This data can also serve as validation sets for emerging photogrammetry methods and form-finding techniques, especially ones involving large deformations and geometric nonlinearities, particularly within the structural engineering field.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    基于下一代测序(NGS)的快速动态实施彻底改变了基因检测,在不久的将来,人类基因组的几乎所有分子改变都可以通过大规模平行测序进行诊断.虽然这一进展将进一步证实人类遗传学在遗传疾病患者的多学科管理中的核心作用,它必须伴随着质量保证措施,以便安全和最佳地使用从基因组诊断中确定的知识。为了实现这一点,已经开发了一些有价值的工具和指南来支持基因组诊断的质量.在本文中,具有基因组分析各个方面经验的作者总结了基因组诊断质量保证的现状,目的是促进该领域核心能力之一的进一步标准化和质量改进。
    The rapid and dynamic implementation of Next-Generation Sequencing (NGS)-based assays has revolutionized genetic testing, and in the near future, nearly all molecular alterations of the human genome will be diagnosable via massive parallel sequencing. While this progress will further corroborate the central role of human genetics in the multidisciplinary management of patients with genetic disorders, it must be accompanied by quality assurance measures in order to allow the safe and optimal use of knowledge ascertained from genome diagnostics. To achieve this, several valuable tools and guidelines have been developed to support the quality of genome diagnostics. In this paper, authors with experience in diverse aspects of genomic analysis summarize the current status of quality assurance in genome diagnostics, with the aim of facilitating further standardization and quality improvement in one of the core competencies of the field.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    偏转轮廓仪用于精确测量同步加速器和X射线自由电子激光器的光束整形光学器件的形式。他们经常利用自动准直仪,其通过评估探测器上的标线图像的位移来测量斜率。根据我们对自动准直仪原始图像数据的特权访问,讨论了通过使用在探测器上不同位置获得的一组重叠图像来减少系统测量误差的新策略。证明了成像特性,例如,例如,几何失真和晕影,可以从这个冗余的图像集提取,而无需求助于外部校准设施。该方法基于以下事实:光罩本身的属性不改变-光罩图像中的所有改变都是由于成像过程。首先,通过结合插值和相关性,可以以最小的误差传播来确定光罩图像相对于参考图像的偏移。其次,分析光罩图像的强度作为其在CCD上的位置的函数,并且计算渐晕校正。第三,分析掩模版图像的尺寸作为其位置的函数,并得出成像失真校正。事实证明,对于自准直仪的不同测量范围和孔径直径,在不求助于外部测量的情况下,可以实现高达4到5倍的系统误差减少。
    Deflectometric profilometers are used to precisely measure the form of beam shaping optics of synchrotrons and X-ray free-electron lasers. They often utilize autocollimators which measure slope by evaluating the displacement of a reticle image on a detector. Based on our privileged access to the raw image data of an autocollimator, novel strategies to reduce the systematic measurement errors by using a set of overlapping images of the reticle obtained at different positions on the detector are discussed. It is demonstrated that imaging properties such as, for example, geometrical distortions and vignetting, can be extracted from this redundant set of images without recourse to external calibration facilities. This approach is based on the fact that the properties of the reticle itself do not change - all changes in the reticle image are due to the imaging process. Firstly, by combining interpolation and correlation, it is possible to determine the shift of a reticle image relative to a reference image with minimal error propagation. Secondly, the intensity of the reticle image is analysed as a function of its position on the CCD and a vignetting correction is calculated. Thirdly, the size of the reticle image is analysed as a function of its position and an imaging distortion correction is derived. It is demonstrated that, for different measurement ranges and aperture diameters of the autocollimator, reductions in the systematic errors of up to a factor of four to five can be achieved without recourse to external measurements.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    在神经监测和解码的交叉点,基于脑电图(EEG)的事件相关电位(ERP)为内在脑功能打开了一个窗口。ERP的稳定性使其在神经科学领域得到了广泛的应用。然而,特定于项目的自定义代码,跟踪用户定义的参数,商业工具的多样性限制了临床应用。
    我们介绍一个开源的,用户友好,和可重复的MATLAB工具箱称为EPAT,包括各种算法的脑电图数据预处理。它提供了基于EEGLAB的模板管道,用于对EEG进行高级多处理,脑磁图,和多导睡眠图数据。参与者评估了EEGLAB和EPAT的14个指标,满意度评分使用Wilcoxon符号秩检验或基于分布正态的配对t检验进行分析。
    EPAT简化了EEG信号浏览和预处理,脑电功率谱分析,独立成分分析,时频分析,ERP波形图,和头皮电压的拓扑分析。用户友好的图形用户界面允许没有编程背景的临床医生和研究人员使用EPAT。
    本文介绍的体系结构,功能,和工具箱的工作流程。EPAT的发布将有助于推进脑电图方法学及其在临床转化研究中的应用。
    UNASSIGNED: At the intersection of neural monitoring and decoding, event-related potential (ERP) based on electroencephalography (EEG) has opened a window into intrinsic brain function. The stability of ERP makes it frequently employed in the field of neuroscience. However, project-specific custom code, tracking of user-defined parameters, and the large diversity of commercial tools have limited clinical application.
    UNASSIGNED: We introduce an open-source, user-friendly, and reproducible MATLAB toolbox named EPAT that includes a variety of algorithms for EEG data preprocessing. It provides EEGLAB-based template pipelines for advanced multi-processing of EEG, magnetoencephalography, and polysomnogram data. Participants evaluated EEGLAB and EPAT across 14 indicators, with satisfaction ratings analyzed using the Wilcoxon signed-rank test or paired t-test based on distribution normality.
    UNASSIGNED: EPAT eases EEG signal browsing and preprocessing, EEG power spectrum analysis, independent component analysis, time-frequency analysis, ERP waveform drawing, and topological analysis of scalp voltage. A user-friendly graphical user interface allows clinicians and researchers with no programming background to use EPAT.
    UNASSIGNED: This article describes the architecture, functionalities, and workflow of the toolbox. The release of EPAT will help advance EEG methodology and its application to clinical translational studies.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

公众号