Programming Languages

编程语言
  • 文章类型: Journal Article
    在过去的几年中,对几乎整体食物概况的分析有了很大的发展。这也导致了更多的数据和获得更多关于食物中健康有益和有害成分的信息的能力。特别是在蛋白质组学领域,软件用于评估,这些都没有为独特的监测问题提供具体方法。使用编程语言Python可以完成额外的和更全面的评估方式。它为质谱数据分析提供了广泛的可能性,但是需要针对特定的功能集进行定制,背后的研究问题。它还提供了各种机器学习方法的适用性。本研究的目的是开发一种从质谱数据中选择和鉴定潜在标记肽的算法。工作流程分为三个步骤:(I)特征工程,(二)化学计量数据分析,和(III)特征识别。第一步是将质谱数据转换为结构,,它可以在Python中应用现有的数据分析包。第二步是用于选择单个特征的数据分析。这些特征在第三步中进一步处理,这是特征识别。在该原理证明方法中示例性使用的数据来自关于热处理对乳蛋白质组/肽组的影响的研究。
    The analysis of almost holistic food profiles has developed considerably over the last years. This has also led to larger amounts of data and the ability to obtain more information about health-beneficial and adverse constituents in food than ever before. Especially in the field of proteomics, software is used for evaluation, and these do not provide specific approaches for unique monitoring questions. An additional and more comprehensive way of evaluation can be done with the programming language Python. It offers broad possibilities by a large ecosystem for mass spectrometric data analysis, but needs to be tailored for specific sets of features, the research questions behind. It also offers the applicability of various machine-learning approaches. The aim of the present study was to develop an algorithm for selecting and identifying potential marker peptides from mass spectrometric data. The workflow is divided into three steps: (I) feature engineering, (II) chemometric data analysis, and (III) feature identification. The first step is the transformation of the mass spectrometric data into a structure, which enables the application of existing data analysis packages in Python. The second step is the data analysis for selecting single features. These features are further processed in the third step, which is the feature identification. The data used exemplarily in this proof-of-principle approach was from a study on the influence of a heat treatment on the milk proteome/peptidome.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    与最近的自动化革命相关的持续技术进步极大地增加了计算机技术在工业中的影响。软件开发和测试是耗时的过程,目前的市场面临着缺乏专业专家的问题。将自动化引入这一领域可以,因此,改善软件工程师的共同工作流程,减少上市时间。尽管已经在基于文本的编程语言中提出了许多代码生成算法,据作者所知,没有一项研究涉及在图形编程环境中实现此类算法,尤其是LabVIEW。由于这一事实,本研究的主要目标是在图形化编程环境LabVIEW中对基于需求的自动代码开发系统进行概念验证。拟议的框架在四个基本基准问题上进行了评估,包含一个字符串模型,一个数字模型,布尔模型和混合型问题模型,它涵盖了基本的编程场景。在所有测试案例中,该算法展示了创建满足所有用户定义要求的功能和无错误解决方案的能力。即使生成的程序带有冗余对象,并且与程序员开发的代码相比要复杂得多,这个事实对代码的执行速度或准确性没有影响。在取得成果的基础上,我们可以得出结论,这项试点研究不仅证明了拟议概念的可行性和可行性,而且在解决线性和二进制编程任务方面也显示出有希望的结果。此外,结果表明,随着进一步的研究,这个探索不足的领域不仅可以成为应用程序开发人员的强大工具,也可以成为非程序员和低技能用户的强大工具。
    Continual technological advances associated with the recent automation revolution have tremendously increased the impact of computer technology in the industry. Software development and testing are time-consuming processes, and the current market faces a lack of specialized experts. Introducing automation to this field could, therefore, improve software engineers\' common workflow and decrease the time to market. Even though many code-generating algorithms have been proposed in textual-based programming languages, to the best of the authors\' knowledge, none of the studies deals with the implementation of such algorithms in graphical programming environments, especially LabVIEW. Due to this fact, the main goal of this study is to conduct a proof-of-concept for a requirement-based automated code-developing system within the graphical programming environment LabVIEW. The proposed framework was evaluated on four basic benchmark problems, encompassing a string model, a numeric model, a boolean model and a mixed-type problem model, which covers fundamental programming scenarios. In all tested cases, the algorithm demonstrated an ability to create satisfying functional and errorless solutions that met all user-defined requirements. Even though the generated programs were burdened with redundant objects and were much more complex compared to programmer-developed codes, this fact has no effect on the code\'s execution speed or accuracy. Based on the achieved results, we can conclude that this pilot study not only proved the feasibility and viability of the proposed concept, but also showed promising results in solving linear and binary programming tasks. Furthermore, the results revealed that with further research, this poorly explored field could become a powerful tool not only for application developers but also for non-programmers and low-skilled users.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    近年来,数据科学方法已经进入医疗保健系统,如放射学,病理学,和放射肿瘤学。在我们的试点研究中,我们开发了一种自动数据挖掘方法,以高速从治疗计划系统(TPS)中提取数据,最大精度,人类互动很少。我们比较了手动数据提取与自动数据挖掘技术所需的时间。
    创建了一个Python编程脚本,以从TPS中提取与患者和治疗有关的指定参数和特征(总共25个特征)。我们成功地实现了数据挖掘的自动化,利用外部束放射治疗设备提供商为接受治疗的整组患者提供的应用程序编程接口环境。
    这个基于Python的内部脚本在0.28±0.03分钟内提取了427名患者的选定特征,以0.04s/plan的惊人速度100%的准确率。相对而言,手动提取25个参数的平均值为4.5±0.33分钟/计划,以及相关的转录和易位错误和缺失的数据信息。这种新方法比传统方法快6850倍。如果我们将提取的特征数量增加一倍,手动特征提取时间将增加近2.5倍,而对于Python脚本,它只增加了1.15倍。
    我们得出的结论是,与手动数据提取相比,我们内部开发的Python脚本可以以更高的速度(>6000倍)从TPS中提取计划数据,并具有最佳的准确性。
    UNASSIGNED: In recent years, data science approaches have entered health-care systems such as radiology, pathology, and radiation oncology. In our pilot study, we developed an automated data mining approach to extract data from a treatment planning system (TPS) with high speed, maximum accuracy, and little human interaction. We compared the amount of time required for manual data extraction versus the automated data mining technique.
    UNASSIGNED: A Python programming script was created to extract specified parameters and features pertaining to patients and treatment (a total of 25 features) from TPS. We successfully implemented automation in data mining, utilizing the application programming interface environment provided by the external beam radiation therapy equipment provider for the whole group of patients who were accepted for treatment.
    UNASSIGNED: This in-house Python-based script extracted selected features for 427 patients in 0.28 ± 0.03 min with 100% accuracy at an astonishing rate of 0.04 s/plan. Comparatively, manual extraction of 25 parameters took an average of 4.5 ± 0.33 min/plan, along with associated transcriptional and transpositional errors and missing data information. This new approach turned out to be 6850 times faster than the conventional approach. Manual feature extraction time increased by a factor of nearly 2.5 if we doubled the number of features extracted, whereas for the Python script, it increased by a factor of just 1.15.
    UNASSIGNED: We conclude that our in-house developed Python script can extract plan data from TPS at a far higher speed (>6000 times) and with the best possible accuracy compared to manual data extraction.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    智能合约(SC)是驻留和运行在区块链上的软件程序。代码可以用不同的语言编写,其共同目的是在托管区块链上实现各种交易。它们由区块链基础设施统治,旨在自动实现传统合同的典型条件。程序必须满足上下文相关的约束,这与传统的软件代码完全不同。特别是,由于字节码是在托管区块链中上传的,尺寸,计算资源,程序的不同部分之间的交互都是有限的。即使特定的编程语言或多或少地实现了与传统语言相同的结构,也是如此:没有与正常软件开发相同的自由。本文使用的工作假设是,智能合约特定的约束应该由特定的软件度量(可能不同于传统的软件度量)捕获。我们在以Solidity编写并上传到以太坊区块链上的85K智能合约上测试了这一假设。我们分析了来自两个存储库“Etherscan”和“智能语料库”的智能合约,并计算了与智能合约相关的一组软件指标的统计数据,并将它们与从更传统的软件项目中提取的指标进行了比较。我们的结果表明,一般来说,智能合约指标比传统软件系统中的相应指标具有更多的限制范围。一些程式化的事实,就像一些度量分布尾部的幂律一样,只是近似的,但是代码行遵循对数正态分布,这使我们想起了传统软件系统中已经发现的相同行为。
    Smart contracts (SC) are software programs that reside and run over a blockchain. The code can be written in different languages with the common purpose of implementing various kinds of transactions onto the hosting blockchain. They are ruled by the blockchain infrastructure with the intent to automatically implement the typical conditions of traditional contracts. Programs must satisfy context-dependent constraints which are quite different from traditional software code. In particular, since the bytecode is uploaded in the hosting blockchain, the size, computational resources, interaction between different parts of the program are all limited. This is true even if the specific programming languages implement more or less the same constructs as that of traditional languages: there is not the same freedom as in normal software development. The working hypothesis used in this article is that Smart Contract specific constraints should be captured by specific software metrics (that may differ from traditional software metrics). We tested this hypothesis on 85K Smart Contracts written in Solidity and uploaded on the Ethereum blockchain. We analyzed Smart Contracts from two repositories \"Etherscan\" and \"Smart Corpus\" and we computed the statistics of a set of software metrics related to Smart Contracts and compared them to the metrics extracted from more traditional software projects. Our results show that generally, Smart Contract metrics have more restricted ranges than the corresponding metrics in traditional software systems. Some of the stylized facts, like power law in the tail of the distribution of some metrics, are only approximate but the lines of code follow a log-normal distribution which reminds us of the same behaviour already found in traditional software systems.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    人们应该假设系统生物学中的计算机模拟实验比它们的湿实验室对应物更不容易受到可重复性问题的影响。因为它们没有自然的生物变异,它们的环境可以完全控制。然而,最近的研究表明,只有一半的已发表的生物系统的数学模型可以复制没有实质性的努力。在本文中,我们以房室结的一维数学模型为例,研究了复制失败或繁琐的潜在原因,我们花了四个月的时间来繁殖。该模型表明,即使是严格的研究,由于缺少信息,也很难重现。方程和参数中的错误,缺乏可用的数据文件,不可执行代码,缺少或不完整的实验方案,缺少方程式背后的基本原理。这些问题中的许多似乎与软件工程中使用单元测试等技术解决的问题相似,回归测试,持续集成,版本控制,档案服务,和一个全面的模块化设计与广泛的文档。应用这些技术,我们使用建模语言Modelica重新实现被检查的模型。生成的工作流程与模型无关,可以转换为SBML,CellML,和其他语言。它通过在物理上与开发环境分离的服务器上的虚拟机中执行自动测试来保证方法的可重复性。此外,它有助于结果的重现性,因为模型更易于理解,并且因为完整的模型代码,实验协议,和仿真数据已发布,并且可以在本文中使用的确切版本中进行访问。我们发现额外的设计和文档工作是合理的,即使只是考虑开发过程中的直接好处,如更容易和更快的调试,增加方程的可理解性,并减少了从文献中查找细节的要求。
    One should assume that in silico experiments in systems biology are less susceptible to reproducibility issues than their wet-lab counterparts, because they are free from natural biological variations and their environment can be fully controlled. However, recent studies show that only half of the published mathematical models of biological systems can be reproduced without substantial effort. In this article we examine the potential causes for failed or cumbersome reproductions in a case study of a one-dimensional mathematical model of the atrioventricular node, which took us four months to reproduce. The model demonstrates that even otherwise rigorous studies can be hard to reproduce due to missing information, errors in equations and parameters, a lack in available data files, non-executable code, missing or incomplete experiment protocols, and missing rationales behind equations. Many of these issues seem similar to problems that have been solved in software engineering using techniques such as unit testing, regression tests, continuous integration, version control, archival services, and a thorough modular design with extensive documentation. Applying these techniques, we reimplement the examined model using the modeling language Modelica. The resulting workflow is independent of the model and can be translated to SBML, CellML, and other languages. It guarantees methods reproducibility by executing automated tests in a virtual machine on a server that is physically separated from the development environment. Additionally, it facilitates results reproducibility, because the model is more understandable and because the complete model code, experiment protocols, and simulation data are published and can be accessed in the exact version that was used in this article. We found the additional design and documentation effort well justified, even just considering the immediate benefits during development such as easier and faster debugging, increased understandability of equations, and a reduced requirement for looking up details from the literature.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    作为重症监护病房(ICU)最重要的技术组成部分之一,当参数偏离正常范围时,通过警报提醒工作人员,连续监测患者的重要参数显著提高了患者的安全性。然而,大量的警报经常使员工不知所措,并可能导致警报疲劳,最近COVID-19加剧了这种情况,并可能危及患者。
    本研究的重点是对ICU患者监测系统的报警数据进行完整且可重复的分析。我们旨在为精通技术的ICU员工开发自己动手(DIY)说明,以自己分析其监测数据,这是开发高效和有效的报警优化策略的基本要素。
    这项观察性研究是使用从2019年21张病床的外科ICU的患者监测系统中提取的警报日志数据进行的。DIY指令在非正式的跨学科小组会议中反复开发。数据分析基于由5个维度组成的框架,每个都有特定的指标:报警负载(例如,每天每张床的警报,报警洪水条件,每个设备和每个临界性的警报),可避免的警报,(eg,技术警报的数量),响应性和警报处理(例如警报持续时间),传感(例如,使用警报暂停功能),和暴露(例如,每个房间类型的警报)。使用R包ggplot2对结果进行可视化,以提供对ICU警报情况的详细了解。
    我们开发了6种DIY指令,应该一步一步迭代地遵循。在收集和分析警报日志数据之前,应(重新)定义警报负载度量。接下来应创建警报指标的直观可视化,并将其呈现给员工,以帮助识别警报数据中的模式,以设计和实施有效的警报管理干预措施。我们提供用于数据准备的脚本和一个R-Markdown文件来创建全面的警报报告。各ICU中的警报负荷通过平均每天每床152.5(SD42.2)警报和警报洪水条件进行量化,平均而言,每天69.55(SD31.12),两者大多发生在早班。大多数警报是由呼吸机发出的,有创血压装置,和心电图(即,高血压和低血压,高呼吸率,低心率)。单人间每天每张床的警报暴露量较高(26%,每张床每天平均172.9/137.2报警)。
    分析ICU警报日志数据可提供对当前警报情况的宝贵见解。我们的结果要求报警管理干预措施,有效地减少报警的数量,以确保患者的安全和ICU工作人员的工作满意度。我们希望我们的DIY说明鼓励其他人遵循分析和发布他们的ICU警报数据。
    As one of the most essential technical components of the intensive care unit (ICU), continuous monitoring of patients\' vital parameters has significantly improved patient safety by alerting staff through an alarm when a parameter deviates from the normal range. However, the vast number of alarms regularly overwhelms staff and may induce alarm fatigue, a condition recently exacerbated by COVID-19 and potentially endangering patients.
    This study focused on providing a complete and repeatable analysis of the alarm data of an ICU\'s patient monitoring system. We aimed to develop do-it-yourself (DIY) instructions for technically versed ICU staff to analyze their monitoring data themselves, which is an essential element for developing efficient and effective alarm optimization strategies.
    This observational study was conducted using alarm log data extracted from the patient monitoring system of a 21-bed surgical ICU in 2019. DIY instructions were iteratively developed in informal interdisciplinary team meetings. The data analysis was grounded in a framework consisting of 5 dimensions, each with specific metrics: alarm load (eg, alarms per bed per day, alarm flood conditions, alarm per device and per criticality), avoidable alarms, (eg, the number of technical alarms), responsiveness and alarm handling (eg alarm duration), sensing (eg, usage of the alarm pause function), and exposure (eg, alarms per room type). Results were visualized using the R package ggplot2 to provide detailed insights into the ICU\'s alarm situation.
    We developed 6 DIY instructions that should be followed iteratively step by step. Alarm load metrics should be (re)defined before alarm log data are collected and analyzed. Intuitive visualizations of the alarm metrics should be created next and presented to staff in order to help identify patterns in the alarm data for designing and implementing effective alarm management interventions. We provide the script we used for the data preparation and an R-Markdown file to create comprehensive alarm reports. The alarm load in the respective ICU was quantified by 152.5 (SD 42.2) alarms per bed per day on average and alarm flood conditions with, on average, 69.55 (SD 31.12) per day that both occurred mostly in the morning shifts. Most alarms were issued by the ventilator, invasive blood pressure device, and electrocardiogram (ie, high and low blood pressure, high respiratory rate, low heart rate). The exposure to alarms per bed per day was higher in single rooms (26%, mean 172.9/137.2 alarms per day per bed).
    Analyzing ICU alarm log data provides valuable insights into the current alarm situation. Our results call for alarm management interventions that effectively reduce the number of alarms in order to ensure patient safety and ICU staff\'s work satisfaction. We hope our DIY instructions encourage others to follow suit in analyzing and publishing their ICU alarm data.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    The aim of this study is to analyze patient movement patterns between hospital departments to derive the underlying intra-hospital movement network, and to assess if movement patterns differ between patients at high or low risk of colonization. For that purpose, we analyzed patient electronic medical record data from five hospitals to extract information on risk stratification and patient intra-hospital movements. Movement patterns were visualized as networks, and network centrality measures were calculated. Next, using an agent-based model where agents represent patients and intra-hospital patient movements were explicitly modeled, we simulated the spread of multidrug resistant enterobacteriacae (MDR-E) inside a hospital. Risk stratification of patients according to certain ICD-10 codes revealed that length of stay, patient age, and mean number of movements per admission were higher in the high-risk groups. Movement networks in all hospitals displayed a high variability among departments concerning their network centrality and connectedness with a few highly connected departments and many weakly connected peripheral departments. Simulating the spread of a pathogen in one hospital network showed positive correlation between department prevalence and network centrality measures. This study highlights the importance of intra-hospital patient movements and their possible impact on pathogen spread. Targeting interventions to departments of higher (weighted) degree may help to control the spread of MDR-E. Moreover, when the colonization status of patients coming from different departments is unknown, a ranking system based on department centralities may be used to design more effective interventions that mitigate pathogen spread.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    近年来,已经提出了使用互联网数据的新形式的综合征监测。开发这些方法是为了帮助早期预测各种病例和疾病的流行病。已经发现,这些系统在监测和预测人口爆发之前是准确的,因此,它们可以用作其他方法的补充。在这项研究中,我们的目标是检查高度传染性疾病,麻疹,因为没有大量的文献使用互联网数据预测麻疹,方法:这项研究是根据欧盟主管当局(欧洲疾病与预防中心-ECDC)提供的5年(2013-2018年)麻疹官方数据以及通过使用Python编码的脚本从Google趋势获得的数据进行的。我们比较了预测五个国家麻疹发展的回归模型。
    结果显示,可以通过Google趋势在时间上对麻疹进行估计和预测,量和整体价差。综合结果显示,麻疹病例与预测病例之间存在很强的相关性(相关系数R=0.779,双尾显著性p<0.01)。综合结果的平均标准误差相对较低,为45.2(12.19%)。然而,麻疹影响相对较低的国家观察到了主要的差异和偏差,比如英国和西班牙。对于这些国家来说,为了改善结果,我们测试了替代模型。
    谷歌趋势对麻疹病例的估计产生了可接受的结果,可以帮助以稳健和合理的方式预测疫情爆发。至少提前2个月。Python脚本可以单独使用,也可以在集成的Internet监视系统的框架内使用,用于跟踪流行病。
    In recent years new forms of syndromic surveillance that use data from the Internet have been proposed. These have been developed to assist the early prediction of epidemics in various cases and diseases. It has been found that these systems are accurate in monitoring and predicting outbreaks before these are observed in population and, therefore, they can be used as a complement to other methods. In this research, our aim is to examine a highly infectious disease, measles, as there is no extensive literature on forecasting measles using Internet data, METHODS: This research has been conducted with official data on measles for 5 years (2013-2018) from the competent authority of the European Union (European Center of Disease and Prevention - ECDC) and data obtained from Google Trends by using scripts coded in Python. We compared regression models forecasting the development of measles in the five countries.
    Results show that measles can be estimated and predicted through Google Trends in terms of time, volume and the overall spread. The combined results reveal a strong relationship of measles cases with the predicted cases (correlation coefficient R= 0.779 in two-tailed significance p< 0.01). The mean standard error was relatively low 45.2 (12.19%) for the combined results. However, major differences and deviations were observed for countries with a relatively low impact of measles, such as the United Kingdom and Spain. For these countries, alternative models were tested in an attempt to improve the results.
    The estimation of measles cases from Google Trends produces acceptable results and can help predict outbreaks in a robust and sound manner, at least 2 months in advance. Python scripts can be used individually or within the framework of an integrated Internet surveillance system for tracking epidemics as the one addressed here.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    To establish the correlation model between Traditional Chinese Medicine (TCM) constitution and physical examination indexes by backpropagation neural network (BPNN) technology. A new method for the identification of TCM constitution in clinics is proposed, which is trying to solve the problem like shortage of TCM doctor, complicated process, low efficiency, and unfavorable application in the current TCM constitution identification methods.
    The corresponding effective samples were formed by sorting out and classifying the original data which were collected from physical examination indexes and TCM constitution types of 950 physical examinees, who were examined at the affiliated hospital of Chengdu University of TCM. The BPNN algorithm was implemented using the C# programming language and Google\'s AI library. Then, the training group and the test (validation) group of the effective samples were, respectively, input into the algorithm, to complete the construction and validation of the target model.
    For all the correlation models built in this paper, the accuracy of the training group and the test group of entire physical examination indexes-constitutional-type network model, respectively, was 88% and 53%, and the error was 0.001. For the other network models, the accuracy of the learning group and the test group and error, respectively, was as follows: liver function (31%, 42%, and 11.7), renal function (41%, 38%, and 6.7), blood routine (56%, 42%, and 2.4), and urine routine (60%, 40%, and 2.6).
    The more the physical examination indexes are used in training, the more accurate the network model is established to predict TCM constitution. The sample data used in this paper showed that there was a relatively strong correlation between TCM constitution and physical examination indexes. Construction of the correlation model between physical examination indexes and TCM constitution is a kind of study for the integration of Chinese and Western medicine, which provides a new approach for the identification of TCM constitution, and it may be expected to avoid the existing problem of TCM constitution identification at present.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Sci-hub)

       PDF(Pubmed)

  • 文章类型: Journal Article
    Programming is one of the most crucial abilities for students in science and technology courses. Few studies on programming ability have considered the effect of students\' construal levels on their learning performance. Therefore, the effects of students\' construal level were explored in this study to fill this research gap and open a new avenue for the improvements in programming ability. The research participants were 110 seventh- and eighth-grade students with basic programming abilities taking an Arduino course. Data were collected from online questionnaires and analyzed using two-way analysis of variance and structural equation modeling to investigate the relationships among construal levels, programming ability, and learning satisfaction. The results revealed that students\' construal levels affect their learning satisfaction and programming ability. These findings indicate that teaching strategies could effectively improve the learning satisfaction and programming ability of junior high school students.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Sci-hub)

       PDF(Pubmed)

公众号