data processing

数据处理
  • 文章类型: Journal Article
    迄今为止,聚和全氟烷基物质(PFAS)对其环境持久性构成了真正的威胁,广泛的物理化学变异性,以及它们的潜在毒性。到目前为止,这些化学物质的很大一部分在结构上仍然未知。这些化学物质,因此,需要使用液相色谱与高分辨率质谱联用(LC-HRMS)实施复杂的非目标分析工作流程,以进行全面的检测和监测。这种方法,尽管全面,并不总是为复杂PFAS混合物的分析提供急需的分析分辨率,例如消防水性成膜泡沫(AFFF)。这项研究巩固了LC×LC技术与高分辨率串联质谱(HRMS/MS)联用的优势,用于鉴定AFFF混合物中的PFAS。在3M和OrchideeAFFF混合物中鉴定出总共57个PFAS同源物系列(HS),这得益于(i)高色谱峰容量(n'2D,c〜300)和(i)通过对HRMS数据的“KendrickMass剩余部分”(RKM)分析提供的质量域分辨率增加。然后,我们试图通过利用可用的参考标准和FluorMatch工作流程与不同氟重复单元的RKM缺陷相结合来注释每个HS的PFAS,如CF2,CF2O,和C2F4O。这种方法产生了12个确定的PFASHS,包括属于全氟烷基羧酸(PFACAs)HS的化合物,全氟烷基磺酸(PFASAs),(N-五氟(5)硫化物)-全氟烷烃磺酸盐(SF5-PFASAs),N-磺丙基二甲基氨丙基全氟烷烃磺酰胺(N-SPAmP-FASA),和N-羧甲基二甲基铵丙基全氟烷烃磺酰胺(N-CMAMP-FASA)。全氟烷基醛和氯化PFASAs的注释类别代表了所研究的AFFF样品中PFASHS的第一个记录。
    To date, poly- and perfluoroalkyl substances (PFAS) represent a real threat for their environmental persistence, wide physicochemical variability, and their potential toxicity. Thus far a large portion of these chemicals remain structurally unknown. These chemicals, therefore, require the implementation of complex non-targeted analysis workflows using liquid chromatography coupled with high-resolution mass spectrometry (LC-HRMS) for their comprehensive detection and monitoring. This approach, even though comprehensive, does not always provide the much-needed analytical resolution for the analysis of complex PFAS mixtures such as fire-fighting aqueous film-forming foams (AFFFs). This study consolidates the advantages of the LC×LC technique hyphenated with high-resolution tandem mass spectrometry (HRMS/MS) for the identification of PFAS in AFFF mixtures. A total of 57 PFAS homolog series (HS) were identified in 3M and Orchidee AFFF mixtures thanks to the (i) high chromatographic peak capacity (n\'2D,c ~ 300) and the (i) increased mass domain resolution provided by the \"remainder of Kendrick Mass\" (RKM) analysis on the HRMS data. Then, we attempted to annotate the PFAS of each HS by exploiting the available reference standards and the FluoroMatch workflow in combination with the RKM defect by different fluorine repeating units, such as CF2, CF2O, and C2F4O. This approach resulted in 12 identified PFAS HS, including compounds belonging to the HS of perfluoroalkyl carboxylic acids (PFACAs), perfluoroalkyl sulfonic acids (PFASAs), (N-pentafluoro(5)sulfide)-perfluoroalkane sulfonates (SF5-PFASAs), N-sulfopropyldimethylammoniopropyl perfluoroalkane sulfonamides (N-SPAmP-FASA), and N-carboxymethyldimethylammoniopropyl perfluoroalkane sulfonamide (N-CMAmP-FASA). The annotated categories of perfluoroalkyl aldehydes and chlorinated PFASAs represent the first record of PFAS HS in the investigated AFFF samples.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    本文提出的研究的主要目的是介绍一种比传统深度神经网络需要更少的计算能力的人工神经网络。该神经网络的开发是通过应用有序模糊数(OFN)实现的。在工业4.0的背景下,有许多应用可以利用该解决方案进行数据处理。它允许在小型设备上的网络边缘部署人工智能,无需将大量数据传输到云服务器进行分析。这样的网络将更容易在小规模解决方案中实现,比如物联网,在未来。本文介绍了对真实系统进行监控的测试结果,异常被检测和预测。
    The primary objective of the research presented in this article is to introduce an artificial neural network that demands less computational power than a conventional deep neural network. The development of this ANN was achieved through the application of Ordered Fuzzy Numbers (OFNs). In the context of Industry 4.0, there are numerous applications where this solution could be utilized for data processing. It allows the deployment of Artificial Intelligence at the network edge on small devices, eliminating the need to transfer large amounts of data to a cloud server for analysis. Such networks will be easier to implement in small-scale solutions, like those for the Internet of Things, in the future. This paper presents test results where a real system was monitored, and anomalies were detected and predicted.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    作为数字表型,从智能手机等消费设备捕获主动和被动数据,变得更加普遍,正确处理数据并从中获得可复制功能的需求变得至关重要。Cortex是用于数字表型数据的开源数据处理管道,针对mindLAMP应用程序的使用进行了优化,全世界近100个研究团队都在使用它。Cortex旨在帮助团队(1)实时评估数字表型数据质量,(2)从数据中得出可复制的临床特征,和(3)实现易于共享的数据可视化。Cortex提供了许多选项来处理数字表型数据,尽管一些常见的方法可能对所有使用它的团队都有价值。本文强调了推理,代码,以及以简化方式充分处理数字表型数据所需的示例步骤。涵盖如何处理数据,评估其质量,派生特征,可视化发现,本文旨在为读者提供适用于分析任何数字表型数据集的知识和技能。更具体地说,本文将向读者传授CortexPython包的来龙去脉。这包括其与mindLAMP平台互动的背景信息,一些基本的命令来学习什么数据可以提取,和更高级的使用软件包与基本的Python混合,目标是创建一个相关矩阵。教程之后,讨论了Cortex的不同用例,连同限制。为了突出临床应用,本文还提供了3种简单的方法来实现在现实世界中使用Cortex的例子。通过了解如何使用数字表型数据并使用Cortex提供可部署的代码,这篇论文旨在展示数字表型的新领域如何既可以被所有人访问,又可以被严格的方法论。
    As digital phenotyping, the capture of active and passive data from consumer devices such as smartphones, becomes more common, the need to properly process the data and derive replicable features from it has become paramount. Cortex is an open-source data processing pipeline for digital phenotyping data, optimized for use with the mindLAMP apps, which is used by nearly 100 research teams across the world. Cortex is designed to help teams (1) assess digital phenotyping data quality in real time, (2) derive replicable clinical features from the data, and (3) enable easy-to-share data visualizations. Cortex offers many options to work with digital phenotyping data, although some common approaches are likely of value to all teams using it. This paper highlights the reasoning, code, and example steps necessary to fully work with digital phenotyping data in a streamlined manner. Covering how to work with the data, assess its quality, derive features, and visualize findings, this paper is designed to offer the reader the knowledge and skills to apply toward analyzing any digital phenotyping data set. More specifically, the paper will teach the reader the ins and outs of the Cortex Python package. This includes background information on its interaction with the mindLAMP platform, some basic commands to learn what data can be pulled and how, and more advanced use of the package mixed with basic Python with the goal of creating a correlation matrix. After the tutorial, different use cases of Cortex are discussed, along with limitations. Toward highlighting clinical applications, this paper also provides 3 easy ways to implement examples of Cortex use in real-world settings. By understanding how to work with digital phenotyping data and providing ready-to-deploy code with Cortex, the paper aims to show how the new field of digital phenotyping can be both accessible to all and rigorous in methodology.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    配电网在发展过程中的长期损耗是由于配电网管理模式落后造成的。传统的配电网网损分析计算方法已不能适应当前配电网的发展环境。为了提高在电力负荷数据中填充缺失值的准确性,提出了粒子群优化算法来优化聚类算法的聚类中心。此外,原始孤立森林异常识别算法可用于检测负荷数据中的异常值,负荷数据的变异系数,提高了算法的识别精度。最后,本文介绍了一种基于广度优先的大数据背景下线损计算方法。以云南省玉溪市配电网系统为例,并进行了仿真实验。结果表明,在部分数据缺失的情况下,增强模糊C均值聚类算法的误差平均为-6.35,标准差为4.015。改进的孤立森林算法受试者在样本异常模糊情况下的特征曲线下面积为0.8586,以最小的下降,根据变异系数,通过分析的细化,发现馈线损失率为7.62%。结果表明,该技术可以快速准确地进行配电网线损分析,可以作为配电网线损管理的指导。
    The long-term loss of distribution network in the process of distribution network development is caused by the backward management mode of distribution network. The traditional analysis and calculation methods of distribution network loss can not adapt to the current development environment of distribution network. To improve the accuracy of filling missing values in power load data, particle swarm optimization algorithm is proposed to optimize the clustering center of the clustering algorithm. Furthermore, the original isolated forest anomaly recognition algorithm can be used to detect outliers in the load data, and the coefficient of variation of the load data is used to improve the recognition accuracy of the algorithm. Finally, this paper introduces a breadth-first-based method for calculating line loss in the context of big data. An example is provided using the distribution network system of Yuxi City in Yunnan Province, and a simulation experiment is carried out. And the findings revealed that the error of the enhanced fuzzy C-mean clustering algorithm was on average - 6.35, with a standard deviation of 4.015 in the situation of partially missing data. The area under the characteristic curve of the improved isolated forest algorithm subjects in the case of the abnormal sample fuzzy situation was 0.8586, with the smallest decrease, based on the coefficient of variation, and through the refinement of the analysis, it was discovered that the feeder line loss rate is 7.62%. It is confirmed that the suggested technique can carry out distribution network line loss analysis fast and accurately and can serve as a guide for managing distribution network line loss.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    随着人们对保护生态系统功能和服务的日益关注,政府制定了公共政策,组织通过其网站免费提供了大量数字数据。另一方面,通过遥感源获取数据以及通过地理信息系统(GIS)和统计工具进行处理的进展,允许前所未有的能力来有效地管理生态系统。然而,这方面的现实世界仍然自相矛盾。原因可能是多种多样的,但是一个强有力的候选人与利益方之间的有限参与有关,这阻碍了所有这些资产的行动。该研究的目的是证明通过将现有的环境政策与环境大数据以及低成本的GIS和数据处理工具相结合,可以显着改善生态系统服务的管理。以位于米纳斯吉拉斯州(巴西)的上RiodasVelhas水文盆地为例,这项研究展示了基于环境变量多样性的主成分分析如何将子流域组装成城市,农业,采矿和异质概况,将生态系统服务的管理指导到最合适的官方制定的保护计划。GIS工具的使用,另一方面,允许将每个计划的实施范围缩小到特定的子盆地。针对许多保护计划,讨论了将优惠管理计划优化分配到优先区域的方法。一个典型的例子是所谓的保护使用潜力(CUP),专门用于保护含水层补给(提供服务)和控制水蚀(调节服务),以及根据土壤能力分配用途(支持服务)。在所有情况下,计划实施准备效率的提高和资源的节约被认为是值得注意的。
    With the growing concerns about the protection of ecosystem functions and services, governments have developed public policies and organizations have produced an awesome volume of digital data freely available through their websites. On the other hand, advances in data acquisition through remote sensed sources and processing through geographic information systems (GIS) and statistical tools, allowed an unprecedent capacity to manage ecosystems efficiently. However, the real-world scenario in that regard remains paradoxically challenging. The reasons can be many and diverse, but a strong candidate relates with the limited engagement among the interest parties that hampers bringing all these assets into action. The aim of the study is to demonstrate that management of ecosystem services can be significantly improved by integrating existing environmental policies with environmental big data and low-cost GIS and data processing tools. Using the Upper Rio das Velhas hydrographic basin located in the state of Minas Gerais (Brazil) as example, the study demonstrated how Principal Components Analysis based on a diversity of environmental variables assembled sub-basins into urban, agriculture, mining and heterogeneous profiles, directing management of ecosystem services to the most appropriate officially established conservation plans. The use of GIS tools, on the other hand, allowed narrowing the implementation of each plan to specific sub-basins. This optimized allocation of preferential management plans to priority areas was discussed for a number of conservation plans. A paradigmatic example was the so-called Conservation Use Potential (CUP) devoted to the protection of aquifer recharge (provision service) and control of water erosion (regulation service), as well as to the allocation of uses as function of soil capability (support service). In all cases, the efficiency gains in readiness for plans\' implementation and economy of resources were prognosed as noteworthy.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    最先进的质谱仪与现代生物信息学算法相结合,用于具有强大统计评分的肽-光谱匹配(PSM),可实现更多可变特征(即,翻译后修饰)从(tandem-)质谱数据中可靠地识别,通常不需要生化浓缩。半特异性蛋白质组搜索,仅在N端或C端进行理论上的酶消化,允许鉴定天然蛋白质末端或由内源性蛋白水解活性产生的那些(也称为“neo-N-termini”分析或“N-terminalomics”)。然而,从这些搜索输出中获得生物学意义在数据挖掘和分析方面可能是具有挑战性的。因此,我们介绍TermineR,一种数据分析方法,用于(1)根据其酶切特异性和已知的蛋白质加工特征对肽进行注释,(2)N端序列模式的丰度差异和富集分析,和(3)新N-终端位置的可视化。我们通过将其应用于多囊肾病小鼠模型的基于串联质量标签(TMT)的蛋白质组学数据来说明TermineR的使用。并评估半特异性搜索对切割事件的生物学解释以及蛋白水解产物对一般蛋白质丰度的可变贡献。TermineR方法和示例数据可在https://github.com/MiguelCos/TermineR上作为R包获得。
    State-of-the-art mass spectrometers combined with modern bioinformatics algorithms for peptide-to-spectrum matching (PSM) with robust statistical scoring allow for more variable features (i.e., post-translational modifications) being reliably identified from (tandem-) mass spectrometry data, often without the need for biochemical enrichment. Semi-specific proteome searches, that enforce a theoretical enzymatic digestion to solely the N- or C-terminal end, allow to identify of native protein termini or those arising from endogenous proteolytic activity (also referred to as \"neo-N-termini\" analysis or \"N-terminomics\"). Nevertheless, deriving biological meaning from these search outputs can be challenging in terms of data mining and analysis. Thus, we introduce TermineR, a data analysis approach for the (1) annotation of peptides according to their enzymatic cleavage specificity and known protein processing features, (2) differential abundance and enrichment analysis of N-terminal sequence patterns, and (3) visualization of neo-N-termini location. We illustrate the use of TermineR by applying it to tandem mass tag (TMT)-based proteomics data of a mouse model of polycystic kidney disease, and assess the semi-specific searches for biological interpretation of cleavage events and the variable contribution of proteolytic products to general protein abundance. The TermineR approach and example data are available as an R package at https://github.com/MiguelCos/TermineR.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    Exposomics旨在测量人类在整个生命周期中的暴露情况以及它们在人体中产生的变化。Exposome规模的研究具有重要的潜力,可以了解环境因素与复杂的多因素疾病之间的相互作用,这些疾病在我们的社会中普遍存在,其起源尚不清楚。在这个框架中,对化学暴露的研究旨在涵盖所有化学暴露及其对人类健康的影响,但是,今天,这个目标似乎仍然不可行,或者至少非常具有挑战性,这使得目前的曝光只是一个概念。此外,化学暴露的研究面临着几个方法学挑战,例如从特定的目标方法转向高通量的多目标和非目标方法,保证生物样品的可用性和质量,以获得高质量的分析数据,应用分析方法的标准化,以及日益复杂的数据集的统计分配,或(非)已知分析物的鉴定。这篇综述从分析的角度讨论了应用曝光概念所涉及的各个步骤。它概述了现有的各种分析方法和仪器,强调它们的互补性,以开发组合分析策略,以推进化学暴露组表征。此外,这篇综述的重点是内分泌干扰化学物质(EDCs),以表明研究即使是一小部分的化学物质暴露是一个巨大的挑战。在暴露组学背景下应用的分析策略已显示出阐明EDC在健康结果中的作用的巨大潜力。然而,将创新方法转化为病因学研究和化学风险评估将需要多学科的努力。与其他专注于曝光组学的评论文章不同,这篇综述从分析化学的角度提供了一个整体的观点,并讨论了整个分析工作流程,以最终获得有价值的结果。
    Exposomics aims to measure human exposures throughout the lifespan and the changes they produce in the human body. Exposome-scale studies have significant potential to understand the interplay of environmental factors with complex multifactorial diseases widespread in our society and whose origin remain unclear. In this framework, the study of the chemical exposome aims to cover all chemical exposures and their effects in human health but, today, this goal still seems unfeasible or at least very challenging, which makes the exposome for now only a concept. Furthermore, the study of the chemical exposome faces several methodological challenges such as moving from specific targeted methodologies towards high-throughput multitargeted and non-targeted approaches, guaranteeing the availability and quality of biological samples to obtain quality analytical data, standardization of applied analytical methodologies, as well as the statistical assignment of increasingly complex datasets, or the identification of (un)known analytes. This review discusses the various steps involved in applying the exposome concept from an analytical perspective. It provides an overview of the wide variety of existing analytical methods and instruments, highlighting their complementarity to develop combined analytical strategies to advance towards the chemical exposome characterization. In addition, this review focuses on endocrine disrupting chemicals (EDCs) to show how studying even a minor part of the chemical exposome represents a great challenge. Analytical strategies applied in an exposomics context have shown great potential to elucidate the role of EDCs in health outcomes. However, translating innovative methods into etiological research and chemical risk assessment will require a multidisciplinary effort. Unlike other review articles focused on exposomics, this review offers a holistic view from the perspective of analytical chemistry and discuss the entire analytical workflow to finally obtain valuable results.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    低温电子显微镜的共同挑战,例如取向偏差,构象多样性,和3D错误分类,复杂的单粒子分析,并导致大量的资源支出。我们之前介绍了一种使用最大费雷特直径分布的计算机模拟方法,费雷特的签名,表征圆盘形样品的样品异质性。这里,我们扩展了Feret签名方法,以确定包含任意形状且仅需要约1000个颗粒的样品的首选方向。该方法使得能够实时调整数据采集参数,以用于优化数据收集策略或帮助决定中断无效成像会话。除了检测首选方向,Feret签名方法可以作为初始图像处理步骤中分类不一致的早期预警系统,一种允许在数据处理中进行战略调整的能力。这些特征将Feret签名确立为在单粒子分析的背景下的有价值的辅助工具。显著加快了结构确定过程。
    Common challenges in cryogenic electron microscopy, such as orientation bias, conformational diversity, and 3D misclassification, complicate single particle analysis and lead to significant resource expenditure. We previously introduced an in silico method using the maximum Feret diameter distribution, the Feret signature, to characterize sample heterogeneity of disc-shaped samples. Here, we expanded the Feret signature methodology to identify preferred orientations of samples containing arbitrary shapes with only about 1000 particles required. This method enables real-time adjustments of data acquisition parameters for optimizing data collection strategies or aiding in decisions to discontinue ineffective imaging sessions. Beyond detecting preferred orientations, the Feret signature approach can serve as an early-warning system for inconsistencies in classification during initial image processing steps, a capability that allows for strategic adjustments in data processing. These features establish the Feret signature as a valuable auxiliary tool in the context of single particle analysis, significantly accelerating the structure determination process.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    智能预测和优化的废水处理厂方法代表了我们管理废水的突破性转变。通过利用数据驱动的预测建模,自动化,和优化策略,它引入了一个全面的框架,旨在提高废水处理操作的效率和可持续性。这种方法包括各种基本阶段,包括数据收集和培训,集成创新的计算模型,如基于黑猩猩的GoogLeNet(CbG),数据处理,和性能预测,同时微调操作参数。所设计的模型是Chimp优化算法和GoogLeNet的混合。GoogLeNet是一种深度卷积架构,黑猩猩优化是基于黑猩猩行为的生物启发优化模型之一。它优化了运行参数,如pH值,剂量率,出水水质,和能源消耗,污水处理厂,修复GoogLeNet中的最佳设置。所设计的模型包括预处理和特征分析等过程,以对运行参数进行有效预测及其优化。值得注意的是,这种创新方法提供了几个关键优势,包括降低运营成本,改善环境结果,更有效的资源管理。通过不断的适应和完善,这种方法不仅优化了污水处理厂的性能,而且有效地应对不断变化的环境挑战,同时节约资源。它代表了在寻求有效和可持续的废水处理实践方面向前迈出的重要一步。RMSE,MAE,地图,建议技术的R2评分分别为1.103、0.233、0.012和0.002。此外,该模型显示,用电量下降到约1.4%,而温室气体排放量比现有技术显著下降到0.12%。
    The intelligent predictive and optimized wastewater treatment plant method represents a ground-breaking shift in how we manage wastewater. By capitalizing on data-driven predictive modeling, automation, and optimization strategies, it introduces a comprehensive framework designed to enhance the efficiency and sustainability of wastewater treatment operations. This methodology encompasses various essential phases, including data gathering and training, the integration of innovative computational models such as Chimp-based GoogLeNet (CbG), data processing, and performance prediction, all while fine-tuning operational parameters. The designed model is a hybrid of the Chimp optimization algorithm and GoogLeNet. The GoogLeNet is a type of deep convolutional architecture, and the Chimp optimization is one of the bio-inspired optimization models based on chimpanzee behavior. It optimizes the operational parameters, such as pH, dosage rate, effluent quality, and energy consumption, of the wastewater treatment plant, by fixing the optimal settings in the GoogLeNet. The designed model includes the process such as pre-processing and feature analysis for the effective prediction of the operation parameters and its optimization. Notably, this innovative approach provides several key advantages, including cost reduction in operations, improved environmental outcomes, and more effective resource management. Through continuous adaptation and refinement, this methodology not only optimizes wastewater treatment plant performance but also effectively tackles evolving environmental challenges while conserving resources. It represents a significant step forward in the quest for efficient and sustainable wastewater treatment practices. The RMSE, MAE, MAPE, and R2 scores for the suggested technique are 1.103, 0.233, 0.012, and 0.002. Also, the model has shown that power usage decreased to about 1.4%, while greenhouse gas emissions have significantly decreased to 0.12% than the existing techniques.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    质谱广泛用于研究各种生物和环境领域的复杂分子机制。启用蛋白质组学等“组学”研究,代谢组学,和脂质组学。随着研究队列越来越大,越来越复杂,有数十到数百个样本,需要强大的质量控制(QC)措施,通过自动化的软件工具变得至关重要,以确保完整性,高质量,以及下游分析的科学结论的有效性,最大限度地减少资源浪费。由于现有的QC工具主要致力于蛋白质组学,需要支持代谢组学的自动化解决方案。为了满足这一需求,我们开发了PeakQC软件,一种独立于组学分子类型的MS数据自动QC工具(即,组学-不可知论者)。它允许自动提取和检查前体离子的峰值度量(例如,质量错误,保留时间,到达时间),并支持各种仪器和采集类型,来自输注实验或使用液相色谱和/或离子迁移谱的前端分离,并且具有/不具有来自数据依赖性或独立采集分析的碎片谱。还生成了碎裂光谱的诊断图。这里,我们使用不同的代表性数据集描述和说明PeakQC的功能,证明其作为提高组学质谱分析质量和可靠性的有价值的工具的实用性。
    Mass spectrometry is broadly employed to study complex molecular mechanisms in various biological and environmental fields, enabling \'omics\' research such as proteomics, metabolomics, and lipidomics. As study cohorts grow larger and more complex with dozens to hundreds of samples, the need for robust quality control (QC) measures through automated software tools becomes paramount to ensure the integrity, high quality, and validity of scientific conclusions from downstream analyses and minimize the waste of resources. Since existing QC tools are mostly dedicated to proteomics, automated solutions supporting metabolomics are needed. To address this need, we developed the software PeakQC, a tool for automated QC of MS data that is independent of omics molecular types (i.e., omics-agnostic). It allows automated extraction and inspection of peak metrics of precursor ions (e.g., errors in mass, retention time, arrival time) and supports various instrumentations and acquisition types, from infusion experiments or using liquid chromatography and/or ion mobility spectrometry front-end separations and with/without fragmentation spectra from data-dependent or independent acquisition analyses. Diagnostic plots for fragmentation spectra are also generated. Here, we describe and illustrate PeakQC\'s functionalities using different representative data sets, demonstrating its utility as a valuable tool for enhancing the quality and reliability of omics mass spectrometry analyses.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

公众号