Data analysis

数据分析
  • 文章类型: Journal Article
    细胞热移位测定(CETSA)使得能够研究细胞环境中的蛋白质-配体相互作用。它提供了有关小分子和大分子配体在相关生理环境中的结合亲和力和特异性的有价值的信息,从而形成了药物发现的独特工具。尽管存在用于扩展CETSA的高通量实验室协议,随后的数据分析和质量控制仍然很费力,限制了实验通量。这里,我们引入了可扩展且稳健的数据分析工作流程,该流程允许将CETSA整合到常规高通量筛查(HT-CETSA)中.这个新的工作流程自动化数据分析,并纳入质量控制(QC),包括异常值检测,样品和板QC,和结果分类。我们描述了工作流程,并展示了它对典型实验工件的鲁棒性,显示缩放效果,并讨论了通过消除手动数据处理步骤来实现数据分析自动化的影响。
    The Cellular Thermal Shift Assay (CETSA) enables the study of protein-ligand interactions in a cellular context. It provides valuable information on the binding affinity and specificity of both small and large molecule ligands in a relevant physiological context, hence forming a unique tool in drug discovery. Though high-throughput lab protocols exist for scaling up CETSA, subsequent data analysis and quality control remain laborious and limit experimental throughput. Here, we introduce a scalable and robust data analysis workflow which allows integration of CETSA into routine high throughput screening (HT-CETSA). This new workflow automates data analysis and incorporates quality control (QC), including outlier detection, sample and plate QC, and result triage. We describe the workflow and show its robustness against typical experimental artifacts, show scaling effects, and discuss the impact of data analysis automation by eliminating manual data processing steps.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    质谱成像(MSI)是表征化合物空间分布的一种有前途的方法。鉴于获取方式的多样化发展和该技术灵敏度的不断提高,生成的数据总量和分析的复杂性都成倍增加,带来越来越多的数据后处理挑战,比如大量的噪音,背景信号干扰,以及由样品位置变化和扫描偏差引起的图像配准偏差,等。深度学习(DL)是广泛用于数据分析和图像重建的强大工具。该工具通过建立和训练神经网络模型,实现数据的自动特征提取,并通过迁移学习实现对目标数据的全面深入的分析,这对于MSI数据分析具有巨大的潜力。本文综述了国内外研究现状,DL在MSI数据分析中的应用进展和挑战,重点关注四个核心阶段:数据预处理、图像重建,聚类分析,和多模态融合。还说明了DL和质谱成像组合在肿瘤诊断和亚型分类研究中的应用。本文还讨论了未来的发展趋势,旨在促进人工智能与质谱技术的更好结合。
    Mass spectrometry imaging (MSI) is a promising method for characterizing the spatial distribution of compounds. Given the diversified development of acquisition methods and continuous improvements in the sensitivity of this technology, both the total amount of generated data and complexity of analysis have exponentially increased, rendering increasing challenges of data postprocessing, such as large amounts of noise, background signal interferences, as well as image registration deviations caused by sample position changes and scan deviations, and etc. Deep learning (DL) is a powerful tool widely used in data analysis and image reconstruction. This tool enables the automatic feature extraction of data by building and training a neural network model, and achieves comprehensive and in-depth analysis of target data through transfer learning, which has great potential for MSI data analysis. This paper reviews the current research status, application progress and challenges of DL in MSI data analysis, focusing on four core stages: data preprocessing, image reconstruction, cluster analysis, and multimodal fusion. The application of a combination of DL and mass spectrometry imaging in the study of tumor diagnosis and subtype classification is also illustrated. This review also discusses trends of development in the future, aiming to promote a better combination of artificial intelligence and mass spectrometry technology.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    本文介绍了从系统文献综述(SLR)中获得的有关使用隐喻和扩展技术进行沉浸式新闻的数据[1]。布尔运算符,英语和西班牙语,用于使用Scopus上的Publish或Perish8软件检索科学文献,2017年至2022年之间的WebofScience和GoogleScholar。在找到所有的科学文献之后,使用选择标准并遵循PRISMA模型进行了方法学过程,获得了总共61篇科学论文的样本.使用DESLOCIS框架对检索到的数据进行评估以及定量和定性分析。第一个数据集[2]包含根据PRISMA语句的阶段检索的出版物的元数据。第二个数据集[3]包含根据DESLOCIS框架的这些出版物的特征。这些数据提供了在沉浸式新闻领域开发新的纵向研究和荟萃分析的可能性。
    This article presents the data obtained from a Systematic Literature Review (SLR) on the use of metaverse and extended technologies for immersive journalism [1]. Boolean operators, both in English and Spanish, were used to retrieve scientific literature using Publish or Perish 8 software on Scopus, Web of Science and Google Scholar between 2017 and 2022. After finding all the scientific literature, a methodological process was carried out using selection criteria and following the PRISMA model to obtain a total sample of 61 scientific articles. The DESLOCIS framework was used for the evaluation and quantitative and qualitative analysis of the retrieved data. The first dataset [2] contains the metadata of the retrieved publications according to the phases of the PRISMA statement. The second dataset [3] contains the characteristics of these publications according to the DESLOCIS framework. The data offer the possibility to develop new longitudinal studies and meta-analyzes in the field of immersive journalism.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    只有验证了所使用的统计方法的基本假设,数据分析才能准确可靠。任何违反这些假设的行为都可能改变分析的结果和结论。在这项研究中,我们开发了智能数据分析V2(SDA-V2),一个交互式和用户友好的Web应用程序,协助统计知识有限的用户进行数据分析,它可以在https://jularatchumnaul自由访问。shinyapps.io/SDA-V2/.SDA-V2自动探索和可视化数据,检查与参数检验相关的基本假设,并为给定数据选择合适的统计方法。此外,SDA-V2可以评估研究仪器的质量,并确定有意义研究所需的最小样本量。然而,虽然SDA-V2是简化统计分析的有价值的工具,它并不能取代对统计原理的基本理解。鼓励研究人员将他们的专业知识与软件的能力相结合,以实现最准确和可信的结果。
    Data analysis can be accurate and reliable only if the underlying assumptions of the used statistical method are validated. Any violations of these assumptions can change the outcomes and conclusions of the analysis. In this study, we developed Smart Data Analysis V2 (SDA-V2), an interactive and user-friendly web application, to assist users with limited statistical knowledge in data analysis, and it can be freely accessed at https://jularatchumnaul.shinyapps.io/SDA-V2/. SDA-V2 automatically explores and visualizes data, examines the underlying assumptions associated with the parametric test, and selects an appropriate statistical method for the given data. Furthermore, SDA-V2 can assess the quality of research instruments and determine the minimum sample size required for a meaningful study. However, while SDA-V2 is a valuable tool for simplifying statistical analysis, it does not replace the need for a fundamental understanding of statistical principles. Researchers are encouraged to combine their expertise with the software\'s capabilities to achieve the most accurate and credible results.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    Esketamine鼻腔喷雾剂(ESK-NS)是一种治疗难治性抑郁症的新药,我们旨在使用美国食品药品监督管理局(FDA)不良事件报告系统(FAERS)数据库在2019年第一季度至2023年第四季度之间检测和表征ESK-NS的不良事件(AE)。报告赔率比(ROR),比例报告比率(PRR),和多项目伽玛泊松收缩器(MGPS)进行检测,以从FAERS数据中检测风险信号,从而确定潜在的ESK-NS-AE关联.共分析了以ESK-NS为主要可疑药物的14,606份AE报告。共有518个首选术语信号和25个系统器官类别,主要集中在精神疾病(33.20%),神经系统疾病(16.67%),一般疾病和给药部位状况(14.21%),其他人得到了。值得注意的是,解离(n=1,093,ROR2,257.80,PRR899.64,EBGM876.86)表现出最高的发生率和信号强度。此外,不常见但明显强烈的AE信号,如手眼协调功能受损,感到内疚,和毫无价值的感觉,被观察到。此外,分离障碍(n=57,ROR510.92,PRR506.70,EBGM386.60)和镇静(n=688,ROR172.68,PRR155.53和EBGM142.05)均表现出强烈的AE信号,前者没有记录在产品特性摘要(SmPC)中。在临床应用中,密切关注精神疾病和神经系统疾病,尤其是分离。同时,临床专业人员应警惕SmPC中未提及的AE信号的发生,并采取预防措施,以确保临床使用的安全性。
    Esketamine nasal spray (ESK-NS) is a new drug for treatment-resistant depression, and we aimed to detect and characterize the adverse events (AEs) of ESK-NS using the Food and Drug Administration (FDA) adverse event reporting system (FAERS) database between 2019 Q1 and 2023 Q4. Reporting odds ratio (ROR), proportional reporting ratio (PRR), and multi-item gamma Poisson shrinker (MGPS) were performed to detect risk signals from the FAERS data to identify potential ESK-NS-AEs associations. A total of 14,606 reports on AEs with ESK-NS as the primary suspected drug were analyzed. A total of 518 preferred terms signals and 25 system organ classes mainly concentrated in psychiatric disorders (33.20%), nervous system disorders (16.67%), general disorders and administration site conditions (14.21%), and others were obtained. Notably, dissociation (n = 1,093, ROR 2,257.80, PRR 899.64, EBGM 876.86) exhibited highest occurrence rates and signal intensity. Moreover, uncommon but significantly strong AEs signals, such as hand-eye coordination impaired, feeling guilty, and feelings of worthlessness, were observed. Additionally, dissociative disorder (n = 57, ROR 510.92, PRR 506.70, EBGM 386.60) and sedation (n = 688, ROR 172.68, PRR 155.53, and EBGM 142.05) both presented strong AE signals, and the former is not recorded in the Summary of Product Characteristics (SmPC). In clinical applications, close attention should be paid to the psychiatric disorders and nervous system disorders, especially dissociation. Meanwhile, clinical professionals should be alert for the occurrence of AEs signals not mentioned in the SmPC and take preventive measures to ensure the safety of clinical use.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    背景:敏感健康数据的广泛和持续重用可以增强人口健康研究对公共决策的作用。本文介绍了支持人口健康信息研究基础设施(PHIRI)的实施和部署的设计原则和不同的构建模块,该方法的优势和挑战以及未来的一些发展。
    方法:PHIRI的设计和实现是基于:(i)数据访问原则-数据不移动但代码移动;(ii)整个工作流程中研究问题的编排,以确保合法,组织,语义和技术互操作性以及(iii)支持四个用例开发的\'master-worker\'联合计算体系结构。
    结果:9个参与者节点和28个Euro-Peristat成员根据预期产出完成了基础设施的部署。因此,每个用例都产生并发布了自己的通用数据模型,分析管道和相应的研究输出。所有数字对象都是根据开放科学和FAIR原则开发和发布的。
    结论:PHIRI以联合方式成功支持了四个用例的开发,克服了重用敏感健康数据的限制,并提供了一种在多个研究节点中实现互操作性的方法。
    BACKGROUND: The extensive and continuous reuse of sensitive health data could enhance the role of population health research on public decisions. This paper describes the design principles and the different building blocks that have supported the implementation and deployment of Population Health Information Research Infrastructure (PHIRI), the strengths and challenges of the approach and some future developments.
    METHODS: The design and implementation of PHIRI have been developed upon: (i) the data visiting principle-data does not move but code moves; (ii) the orchestration of the research question throughout a workflow that ensured legal, organizational, semantic and technological interoperability and (iii) a \'master-worker\' federated computational architecture that supported the development of four uses cases.
    RESULTS: Nine participants nodes and 28 Euro-Peristat members completed the deployment of the infrastructure according to the expected outputs. As a consequence, each use case produced and published their own common data model, the analytical pipeline and the corresponding research outputs. All the digital objects were developed and published according to Open Science and FAIR principles.
    CONCLUSIONS: PHIRI has successfully supported the development of four use cases in a federated manner, overcoming limitations for the reuse of sensitive health data and providing a methodology to achieve interoperability in multiple research nodes.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Systematic Review
    背景:基于文本的数字媒体平台彻底改变了通信和信息共享,在心理健康和自杀预防领域提供宝贵的知识和理解。
    目的:本系统综述旨在确定如何将机器学习和数据分析应用于基于文本的数字媒体数据,以了解心理健康并帮助预防自杀。
    方法:对来自以下主要电子数据库的研究论文进行了系统综述:WebofScience,MEDLINE,Embase(通过MEDLINE),和PsycINFO(通过MEDLINE)。使用GoogleScholar进行手动搜索,以补充数据库搜索。
    结果:总体而言,包括19项研究,关于如何应用数据分析和机器学习技术的五个主要主题:(1)作为个人心理健康的预测指标,(2)了解个人心理健康和自杀行为是如何沟通的,(3)检测精神障碍和自杀风险,(4)确定寻求帮助的心理健康困难,(5)确定支持心理健康的干预措施的有效性。
    结论:我们的研究结果表明,数据分析和机器学习可用于获得有价值的见解,例如:与抑郁症有关的基于网络的对话在不同种族之间有所不同,青少年比成年人更频繁地进行关于自杀的网络对话,在基于网络的心理健康社区寻求支持的人在获得在线支持后感觉更好。数字工具和心理健康应用程序正在成功用于管理心理健康,特别是通过COVID-19的流行,在此期间,分析显示焦虑和抑郁增加,基于网络的社区在减少大流行期间的孤立方面发挥了作用。预测分析也被证明具有潜力,虚拟现实在提供预防性或治疗性护理方面显示出有希望的结果。未来的研究工作可以集中在优化算法上,以增强基于文本的数字媒体分析在心理健康和自杀预防方面的潜力。在解决抑郁症时,关键的一步是确定导致幸福的因素,并使用机器学习来预测这些幸福的来源。这可以扩展到理解各种活动如何在不同的社会经济群体中提高幸福感。利用从这些数据分析和机器学习中收集的见解,有机会制定数字干预措施,比如聊天机器人,旨在提供支持和解决心理健康挑战和自杀预防。
    BACKGROUND: Text-based digital media platforms have revolutionized communication and information sharing, providing valuable access to knowledge and understanding in the fields of mental health and suicide prevention.
    OBJECTIVE: This systematic review aimed to determine how machine learning and data analysis can be applied to text-based digital media data to understand mental health and aid suicide prevention.
    METHODS: A systematic review of research papers from the following major electronic databases was conducted: Web of Science, MEDLINE, Embase (via MEDLINE), and PsycINFO (via MEDLINE). The database search was supplemented by a hand search using Google Scholar.
    RESULTS: Overall, 19 studies were included, with five major themes as to how data analysis and machine learning techniques could be applied: (1) as predictors of personal mental health, (2) to understand how personal mental health and suicidal behavior are communicated, (3) to detect mental disorders and suicidal risk, (4) to identify help seeking for mental health difficulties, and (5) to determine the efficacy of interventions to support mental well-being.
    CONCLUSIONS: Our findings show that data analysis and machine learning can be used to gain valuable insights, such as the following: web-based conversations relating to depression vary among different ethnic groups, teenagers engage in a web-based conversation about suicide more often than adults, and people seeking support in web-based mental health communities feel better after receiving online support. Digital tools and mental health apps are being used successfully to manage mental health, particularly through the COVID-19 epidemic, during which analysis has revealed that there was increased anxiety and depression, and web-based communities played a part in reducing isolation during the pandemic. Predictive analytics were also shown to have potential, and virtual reality shows promising results in the delivery of preventive or curative care. Future research efforts could center on optimizing algorithms to enhance the potential of text-based digital media analysis in mental health and suicide prevention. In addressing depression, a crucial step involves identifying the factors that contribute to happiness and using machine learning to forecast these sources of happiness. This could extend to understanding how various activities result in improved happiness across different socioeconomic groups. Using insights gathered from such data analysis and machine learning, there is an opportunity to craft digital interventions, such as chatbots, designed to provide support and address mental health challenges and suicide prevention.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    背景:临床研究通常受到可用资源的限制,这导致样本量受到限制。当样本量太小时,我们使用模拟数据来说明研究意义。
    结果:使用2个理论种群,每个种群N=1000,我们从每个种群中随机抽样10个,并进行统计比较,以帮助得出两个种群是否不同的结论。这个练习重复了总共4项研究:2个得出的结论是,这2个群体在统计学上有显著差异,而2则无统计学差异。
    结论:我们的模拟例子表明,样本量在临床研究中起着重要作用。结果和结论,就手段估计而言,中位数,皮尔逊相关性,卡方检验,和P值,是不可靠的小样本。
    BACKGROUND: Clinical studies are often limited by resources available, which results in constraints on sample size. We use simulated data to illustrate study implications when the sample size is too small.
    RESULTS: Using 2 theoretical populations each with N = 1000, we randomly sample 10 from each population and conduct a statistical comparison, to help make a conclusion about whether the 2 populations are different. This exercise is repeated for a total of 4 studies: 2 concluded that the 2 populations are statistically significantly different, while 2 showed no statistically significant difference.
    CONCLUSIONS: Our simulated examples demonstrate that sample sizes play important roles in clinical research. The results and conclusions, in terms of estimates of means, medians, Pearson correlations, chi-square test, and P values, are unreliable with small samples.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    背景和目的:尽管他汀类药物被推荐用于急性缺血性卒中的二级预防,一些基于人群的研究和临床证据表明,他们可能会增加颅内出血的风险.在这个嵌套的病例对照研究中,我们使用台湾的全国全民健康保险数据库,调查了在台湾急性缺血性卒中患者的他汀类药物治疗与他们随后的脑出血风险和全因死亡率之间的可能关联.材料和方法:所有数据均来自台湾国家健康保险研究数据库。急性缺血性卒中患者分为接受他汀类药物药物治疗的队列和未接受他汀类药物药物治疗的对照组。1:1匹配年龄,性别,和索引日,并进行了倾向评分匹配,产生39366例病例和39366例对照。主要结果是长期的后续脑出血和全因死亡率。使用Fine和Gray回归风险模型估算了随后的脑出血与全因死亡率之间的竞争风险。结果:急性缺血性卒中后接受他汀类药物药物治疗的患者,其后续脑出血风险显著降低(p<0.0001),全因死亡率降低(p<0.0001)。Low,中度,高剂量他汀类药物与随后脑出血(调整sHRs0.82,0.74,0.53)和全因死亡率(调整sHRs0.75,0.74,0.74)的风险显著降低相关,分别。结论:发现他汀类药物药物治疗可安全有效地降低台湾急性缺血性卒中患者随后的脑出血和全因死亡率的风险。
    Background and Objectives: Although statins are recommended for secondary prevention of acute ischemic stroke, some population-based studies and clinical evidence suggest that they might be used with an increased risk of intracranial hemorrhage. In this nested case-control study, we used Taiwan\'s nationwide universal health insurance database to investigate the possible association between statin therapy prescribed to acute ischemic stroke patients and their risk of subsequent intracerebral hemorrhage and all-cause mortality in Taiwan. Materials and Methods: All data were retrospectively obtained from Taiwan\'s National Health Insurance Research Database. Acute ischemic stroke patients were divided into a cohort receiving statin pharmacotherapy and a control cohort not receiving statin pharmacotherapy. A 1:1 matching for age, gender, and index day, and propensity score matching was conducted, producing 39,366 cases and 39,366 controls. The primary outcomes were long-term subsequent intracerebral hemorrhage and all-cause mortality. The competing risk between subsequent intracerebral hemorrhage and all-cause mortality was estimated using the Fine and Gray regression hazards model. Results: Patients receiving statin pharmacotherapy after an acute ischemic stroke had a significantly lower risk of subsequent intracerebral hemorrhage (p < 0.0001) and lower all-cause mortality rates (p < 0.0001). Low, moderate, and high dosages of statin were associated with significantly decreased risks for subsequent intracerebral hemorrhage (adjusted sHRs 0.82, 0.74, 0.53) and all-cause mortality (adjusted sHRs 0.75, 0.74, 0.74), respectively. Conclusions: Statin pharmacotherapy was found to safely and effectively reduce the risk of subsequent intracerebral hemorrhage and all-cause mortality in acute ischemic stroke patients in Taiwan.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    在社会经济不断发展的背景下,人们越来越关注食品质量和安全。人们越来越意识到健康饮食对身体健康的至关重要性;因此,检测食品污染的需求持续增长。同时,全球食品贸易的迅速扩张使人们对高品质食品的追求更加迫切。然而,传统的食品分析方法有一定的局限性,主要表现在对食品质量评价高度依赖个人主观判断。在这种情况下,人工智能和生物传感器的出现为食品质量评价提供了新的可能性。本文提出了一种综合方法,包括汇总与食品质量指数相关的数据,并开发相应的评估模型,以突出人工智能和生物传感器在食品质量评估中的有效性和全面性。全面讨论了该方法在食品安全领域的潜在前景和挑战,旨在为今后的研究和实践提供有价值的参考。
    Against the backdrop of continuous socio-economic development, there is a growing concern among people about food quality and safety. Individuals are increasingly realizing the critical importance of healthy eating for bodily health; hence the continuous rise in demand for detecting food pollution. Simultaneously, the rapid expansion of global food trade has made people\'s pursuit of high-quality food more urgent. However, traditional methods of food analysis have certain limitations, mainly manifested in the high degree of reliance on personal subjective judgment for assessing food quality. In this context, the emergence of artificial intelligence and biosensors has provided new possibilities for the evaluation of food quality. This paper proposes a comprehensive approach that involves aggregating data relevant to food quality indices and developing corresponding evaluation models to highlight the effectiveness and comprehensiveness of artificial intelligence and biosensors in food quality evaluation. The potential prospects and challenges of this method in the field of food safety are comprehensively discussed, aiming to provide valuable references for future research and practice.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

公众号