Reinforcement Learning

强化学习
  • 文章类型: Journal Article
    背景:当前的管理屏幕检测到的肺结节的指南提供了基于规则的建议,以立即进行诊断检查或每隔3、6或12个月进行随访。缺乏定制的访问计划。
    目的:使用强化学习(RL)制定个性化的筛查计划,并评估基于RL的政策模型的有效性。
    方法:使用嵌套的案例控制设计,我们回顾性地确定了308例癌症患者,这些患者在国家肺癌筛查试验的至少两轮筛查中筛查结果为阳性.我们建立了一个对照组,包括没有癌症的结节患者,根据癌症诊断年份匹配(1:1)。通过生成10,164个序列决策事件,我们训练了基于RL的策略模型,仅包含结节直径,结合结节外观(衰减和边缘)和/或患者信息(年龄,性别,吸烟状况,包年,和家族史)。我们计算了误诊率,漏诊,和延迟诊断,并比较了基于RL的政策模型和基于规则的随访协议(国家综合癌症网络指南;中国肺癌筛查和早期检测指南)的性能。
    结果:我们确定了某些变量之间的显着相互作用(例如,结节形状和患者吸烟包年,超出指南协议中考虑的范围)和后续测试间隔的选择,从而影响决策序列的质量。在验证中,一个基于RL的政策模型的误诊率为12.3%,9.7%为漏诊,延迟诊断为11.7%。与两种基于规则的协议相比,三个性能最佳的基于RL的策略模型一致地证明了基于疾病特征(良性或恶性)的特定患者亚组的最佳性能,结节表型(大小,形状,和衰减),和个人属性。
    结论:这项研究强调了使用基于RL的方法的潜力,该方法在临床上可解释且性能稳健,以开发个性化的肺癌筛查时间表。我们的发现为增强当前的癌症筛查系统提供了机会。
    BACKGROUND: The current guidelines for managing screen-detected pulmonary nodules offer rule-based recommendations for immediate diagnostic work-up or follow-up at intervals of 3, 6, or 12 months. Customized visit plans are lacking.
    OBJECTIVE: To develop individualized screening schedules using reinforcement learning (RL) and evaluate the effectiveness of RL-based policy models.
    METHODS: Using a nested case-control design, we retrospectively identified 308 patients with cancer who had positive screening results in at least two screening rounds in the National Lung Screening Trial. We established a control group that included cancer-free patients with nodules, matched (1:1) according to the year of cancer diagnosis. By generating 10,164 sequence decision episodes, we trained RL-based policy models, incorporating nodule diameter alone, combined with nodule appearance (attenuation and margin) and/or patient information (age, sex, smoking status, pack-years, and family history). We calculated rates of misdiagnosis, missed diagnosis, and delayed diagnosis, and compared the performance of RL-based policy models with rule-based follow-up protocols (National Comprehensive Cancer Network guideline; China Guideline for the Screening and Early Detection of Lung Cancer).
    RESULTS: We identified significant interactions between certain variables (e.g., nodule shape and patient smoking pack-years, beyond those considered in guideline protocols) and the selection of follow-up testing intervals, thereby impacting the quality of the decision sequence. In validation, one RL-based policy model achieved rates of 12.3% for misdiagnosis, 9.7% for missed diagnosis, and 11.7% for delayed diagnosis. Compared with the two rule-based protocols, the three best-performing RL-based policy models consistently demonstrated optimal performance for specific patient subgroups based on disease characteristics (benign or malignant), nodule phenotypes (size, shape, and attenuation), and individual attributes.
    CONCLUSIONS: This study highlights the potential of using an RL-based approach that is both clinically interpretable and performance-robust to develop personalized lung cancer screening schedules. Our findings present opportunities for enhancing the current cancer screening system.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    患者之间的治疗反应变异性是临床实践中的常见现象。对于许多药物,这种个体间的变异性不需要太多的(如果有的话)个体化给药策略。然而,对于一些药物,包括化疗和一些单克隆抗体治疗,需要个体化剂量以避免有害不良事件。基于模型的精确给药(MIPD)是一种新兴的方法,用于指导其他难以给药的药物的给药方案的个性化。已经提出了几种MIPD方法来预测给药策略,包括回归,强化学习(RL)和药代动力学和药效学(PKPD)模型。缺少一个统一的框架来研究这些方法的优点和局限性。我们开发了一个框架来模拟临床MIPD试验,提供一种成本和时间有效的方法来测试不同的MIPD方法。我们框架的核心是一个临床试验模型,它模拟了临床实践中挑战成功治疗个体化的复杂性。我们使用华法林治疗作为用例演示了该框架,并研究了三种流行的MIPD方法:1.神经网络回归;2.深RL;和3。PKPD建模。我们发现PKPD模型使华法林给药方案具有最高的成功率和最高的效率:75.1%的个体在模拟试验结束时显示INR在治疗范围内;治疗范围(TTR)的中位时间为74%。相比之下,回归模型和深度RL模型的成功率分别为47.0%和65.8%,TTRs中位数为45%和68%。我们还发现MIPD模型可以实现不同程度的个性化:回归模型将给药方案个性化,直至由协变量解释的可变性;DeepRL模型和PKPD模型将给药方案个性化,也考虑了使用监测数据的额外变化。然而,深度RL模型侧重于治疗反应的控制,而PKPD模型也使用数据来进一步个性化预测。
    Treatment response variability across patients is a common phenomenon in clinical practice. For many drugs this inter-individual variability does not require much (if any) individualisation of dosing strategies. However, for some drugs, including chemotherapies and some monoclonal antibody treatments, individualisation of dosages are needed to avoid harmful adverse events. Model-informed precision dosing (MIPD) is an emerging approach to guide the individualisation of dosing regimens of otherwise difficult-to-administer drugs. Several MIPD approaches have been suggested to predict dosing strategies, including regression, reinforcement learning (RL) and pharmacokinetic and pharmacodynamic (PKPD) modelling. A unified framework to study the strengths and limitations of these approaches is missing. We develop a framework to simulate clinical MIPD trials, providing a cost and time efficient way to test different MIPD approaches. Central for our framework is a clinical trial model that emulates the complexities in clinical practice that challenge successful treatment individualisation. We demonstrate this framework using warfarin treatment as a use case and investigate three popular MIPD methods: 1. Neural network regression; 2. Deep RL; and 3. PKPD modelling. We find that the PKPD model individualises warfarin dosing regimens with the highest success rate and the highest efficiency: 75.1% of the individuals display INRs inside the therapeutic range at the end of the simulated trial; and the median time in the therapeutic range (TTR) is 74%. In comparison, the regression model and the deep RL model have success rates of 47.0% and 65.8%, and median TTRs of 45% and 68%. We also find that the MIPD models can attain different degrees of individualisation: the Regression model individualises dosing regimens up to variability explained by covariates; the Deep RL model and the PKPD model individualise dosing regimens accounting also for additional variation using monitoring data. However, the Deep RL model focusses on control of the treatment response, while the PKPD model uses the data also to further the individualisation of predictions.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    背景:灵活的自适应算法监测测试(FAAST)计划代表了一种改进新的传染病病例检测的创新方法;它在这里被部署用于筛查和诊断SARS-CoV-2。随着COVID-19治疗的出现,发现感染SARS-CoV-2的个体是临床和公共卫生的当务之急。虽然这些类型的贝叶斯搜索算法在其他设置中被广泛使用(例如,寻找被击落的飞机,在潜艇回收中,并协助石油勘探),这是该领域首次将贝叶斯自适应方法用于主动疾病监测.
    目的:这项研究的目的是评估贝叶斯搜索算法,以针对社区中SARS-CoV-2传播的热点,目的是随着时间的推移,在哥伦布的多个地点发现大多数病例。俄亥俄州,2021年8月至10月。
    方法:用于此项目的直接弹出式SARS-CoV-2测试的算法基于汤普森采样,其中的目的是根据每个测试地点的先验概率分布的采样,最大限度地提高一组测试地点中诊断出的SARS-CoV-2新病例的平均数量。耶鲁大学之间的学术-政府伙伴关系,俄亥俄州立大学,威克森林大学,俄亥俄州卫生部,俄亥俄州国民警卫队,哥伦布大都会图书馆对土匪算法进行了研究,以最大限度地检测2021年俄亥俄州城市的SARS-CoV-2新病例。该计划在哥伦布的13个地点建立了弹出式COVID-19测试站点,包括图书馆分支机构,娱乐和社区中心,电影院,无家可归者收容所,家庭服务中心,和社区活动网站。我们的团队在16个测试事件中进行了0到56个测试,每个事件进行的总体平均值为25.3,移动平均值随时间增加。向那些接近弹出式网站以鼓励他们参与的人提供了小型激励措施,包括礼品卡和带回家的快速抗原测试。
    结果:随着时间的推移,正如预期的那样,贝叶斯搜索算法将测试工作定向到新诊断结果较高的位置.令人惊讶的是,该算法的使用还最大限度地提高了服务不足社区的少数民族居民的案件识别能力,尤其是非洲裔美国人,相对于测试地点所在的当地邮政编码的人口统计概况,参与者人数过多。
    结论:这项研究表明,使用bandit算法的弹出式测试策略可以在大流行期间在城市环境中可行地部署。这是首次在现实世界中使用这些算法进行疾病监测,并且代表了评估其使用有效性的关键步骤,以最大程度地检测未确诊的SARS-CoV-2和其他感染病例。比如HIV。
    The Flexible Adaptive Algorithmic Surveillance Testing (FAAST) program represents an innovative approach for improving the detection of new cases of infectious disease; it is deployed here to screen and diagnose SARS-CoV-2. With the advent of treatment for COVID-19, finding individuals infected with SARS-CoV-2 is an urgent clinical and public health priority. While these kinds of Bayesian search algorithms are used widely in other settings (eg, to find downed aircraft, in submarine recovery, and to aid in oil exploration), this is the first time that Bayesian adaptive approaches have been used for active disease surveillance in the field.
    This study\'s objective was to evaluate a Bayesian search algorithm to target hotspots of SARS-CoV-2 transmission in the community with the goal of detecting the most cases over time across multiple locations in Columbus, Ohio, from August to October 2021.
    The algorithm used to direct pop-up SARS-CoV-2 testing for this project is based on Thompson sampling, in which the aim is to maximize the average number of new cases of SARS-CoV-2 diagnosed among a set of testing locations based on sampling from prior probability distributions for each testing site. An academic-governmental partnership between Yale University, The Ohio State University, Wake Forest University, the Ohio Department of Health, the Ohio National Guard, and the Columbus Metropolitan Libraries conducted a study of bandit algorithms to maximize the detection of new cases of SARS-CoV-2 in this Ohio city in 2021. The initiative established pop-up COVID-19 testing sites at 13 Columbus locations, including library branches, recreational and community centers, movie theaters, homeless shelters, family services centers, and community event sites. Our team conducted between 0 and 56 tests at the 16 testing events, with an overall average of 25.3 tests conducted per event and a moving average that increased over time. Small incentives-including gift cards and take-home rapid antigen tests-were offered to those who approached the pop-up sites to encourage their participation.
    Over time, as expected, the Bayesian search algorithm directed testing efforts to locations with higher yields of new diagnoses. Surprisingly, the use of the algorithm also maximized the identification of cases among minority residents of underserved communities, particularly African Americans, with the pool of participants overrepresenting these people relative to the demographic profile of the local zip code in which testing sites were located.
    This study demonstrated that a pop-up testing strategy using a bandit algorithm can be feasibly deployed in an urban setting during a pandemic. It is the first real-world use of these kinds of algorithms for disease surveillance and represents a key step in evaluating the effectiveness of their use in maximizing the detection of undiagnosed cases of SARS-CoV-2 and other infections, such as HIV.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    深度学习在量化风险管理领域的广泛应用仍然是一个相对较新的现象。本文介绍了深度资产负债管理(“DeepALM”)的关键概念,用于整个期限结构中资产和负债管理的技术改造。该方法对广泛的应用产生了深远的影响,例如财务主管的最佳决策,商品的优化采购或水力发电厂的优化。作为副产品,基于目标的投资或资产负债管理(ALM)的抽象方面与我们社会的紧迫挑战有关。我们在程式化的情况下说明了该方法的潜力。
    The extensive application of deep learning in the field of quantitative risk management is still a relatively recent phenomenon. This article presents the key notions of Deep Asset-Liability-Management (\"Deep ALM\") for a technological transformation in the management of assets and liabilities along a whole term structure. The approach has a profound impact on a wide range of applications such as optimal decision making for treasurers, optimal procurement of commodities or the optimization of hydroelectric power plants. As a by-product, intriguing aspects of goal-based investing or Asset-Liability-Management (ALM) in abstract terms concerning urgent challenges of our society are expected alongside. We illustrate the potential of the approach in a stylized case.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    在流程工业智能制造的背景下,传统的基于模型的优化控制方法不能适应工况或运行模式急剧变化的情况。强化学习(RL)通过与环境的交互直接实现控制目标,并且在存在不确定性的情况下具有显着的优势,因为它不需要操作工厂的显式模型。然而,大多数RL算法在存在模式变化的情况下无法保留迁移学习能力,这成为工业过程控制应用的实际障碍。为了解决这些问题,我们设计了一个框架,使用本地数据增强来提高训练效率和迁移学习(适应性)性能。因此,本文提出了一种新的RL控制算法,CBR-MA-DDPG,有机地集成基于案例的推理(CBR),模型辅助(MA)体验增强,和深度确定性政策梯度(DDPG)。当操作模式改变时,CBR-MA-DDPG可以快速适应变化的环境,并在几个训练事件中实现所需的控制性能。对连续搅拌釜反应器(CSTR)和有机朗肯循环(ORC)的实验分析证明了该方法在适应性和控制性能/鲁棒性方面的优越性。结果表明,CBR-MA-DDPG代理的控制性能优于常规PI和MPC控制方案,它比最先进的DDPG具有更高的训练效率,具有模式转换情况的迁移学习场景中的TD3和PPO算法。
    In the context of intelligent manufacturing in the process industry, traditional model-based optimization control methods cannot adapt to the situation of drastic changes in working conditions or operating modes. Reinforcement learning (RL) directly achieves the control objective by interacting with the environment, and has significant advantages in the presence of uncertainty since it does not require an explicit model of the operating plant. However, most RL algorithms fail to retain transfer learning capabilities in the presence of mode variation, which becomes a practical obstacle to industrial process control applications. To address these issues, we design a framework that uses local data augmentation to improve the training efficiency and transfer learning (adaptability) performance. Therefore, this paper proposes a novel RL control algorithm, CBR-MA-DDPG, organically integrating case-based reasoning (CBR), model-assisted (MA) experience augmentation, and deep deterministic policy gradient (DDPG). When the operating mode changes, CBR-MA-DDPG can quickly adapt to the varying environment and achieve the desired control performance within several training episodes. Experimental analyses on a continuous stirred tank reactor (CSTR) and an organic Rankine cycle (ORC) demonstrate the superiority of the proposed method in terms of both adaptability and control performance/robustness. The results show that the control performance of the CBR-MA-DDPG agent outperforms the conventional PI and MPC control schemes, and that it has higher training efficiency than the state-of-the-art DDPG, TD3, and PPO algorithms in transfer learning scenarios with mode shift situations.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    病例日志是外科教育的基础数据,然而,病例一直被低估。测井行为是由多种人和系统因素驱动的,包括时间限制,便于案例数据检索,访问数据输入工具,和程序代码决策工具。
    我们研究了三个中等规模的病例记录趋势,普外科培训项目分别为2016年9月-2020年10月、2019年1月-2020年10月和2019年5月-2020年10月。在整个计划中,我们比较了居民直接登录ACGME与通过基于机器学习的案例记录辅助工具的居民教育平台每周记录的案例数量。我们检查了4个连续阶段的案例记录模式:平台访问之前的基线默认ACGME日志记录(P0“手动”),全平台日志记录辅助(P1“辅助”),需要手动输入数据而不进行数据集成的部分平台辅助(P2"Notebook"),并通过日志记录帮助恢复完全集成的平台(P3“已恢复”)。
    自2016年以来,这3个项目的171名居民利用该平台记录了31,385个案例。智能案例记录辅助显着提高了案例记录率,从P0中手动输入的1.44±1.48例到P1中每个居民每周通过平台输入的4.77±2.45例(p值<0.00001)。尽管在平台的数据连接暂停时需要手动输入数据,该工具有助于将ACGME的总体病例记录增加至每周2.85±2.37例(p值=0.0002).恢复数据连接后,通过该平台,病例记录水平上升到每周4.54±3.33例,相当于P1水平(差异不显著,p值=0.57)。
    在高质量的病例日志中绘制系统和人为因素的影响,使我们能够针对干预措施,以不断改进对外科住院医师的培训。系统级因素,例如访问替代自动化驱动工具和操作时间表集成平台以协助ACGME病例日志,对日志中捕获的病例数量有重大影响。
    Case logs are foundational data in surgical education, yet cases are consistently under-reported. Logging behavior is driven by multiple human and systems factors, including time constraints, ease of case data retrieval, access to data-entry tools, and procedural code decision tools.
    We examined case logging trends at three mid-sized, general surgery training programs from September 2016-October 2020, January 2019-October 2020 and May 2019-October 2020, respectively. Across the programs we compared the number of cases logged per week when residents logged directly to ACGME versus via a resident education platform with machine learning-based case logging assistance tools. We examined case logging patterns across 4 consecutive phases: baseline default ACGME logging prior to platform access (P0 \"Manual\"), full platform logging assistance (P1 \"Assisted\"), partial platform assistance requiring manual data entry without data integrations (P2 \"Notebook\"), and resumed fully integrated platform with logging assistance (P3 \"Resumed\").
    31,385 cases were logged utilizing the platform since 2016 by 171 residents across the 3 programs.Intelligent case logging assistance significantly increased case logging rates, from 1.44 ± 1.48 cases by manual entry in P0 to 4.77 ± 2.45 cases per resident per week via the platform in P1 (p-value < 0.00001). Despite the burden of manual data entry when the platform\'s data connectivity was paused, the tool helped to increase overall case logging into ACGME to 2.85 ± 2.37 cases per week (p-value = 0.0002). Upon resuming the data connectivity, case logging levels rose to 4.54 ± 3.33 cases per week via the platform, equivalent to P1 levels (insignificant difference, p-value = 0.57).
    Mapping the influence of systems and human factors in high-quality case logs allows us to target interventions to continually improve the training of surgical residents. System level factors such as access to alternate automation-drive tools and operative schedule integrated platforms to assist in ACGME case log has a significant impact on the number of cases captured in logs.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    A problem is a situation in which an agent seeks to attain a given goal without knowing how to achieve it. Human problem solving is typically studied as a search in a problem space composed of states (information about the environment) and operators (to move between states). A problem such as playing a game of chess has 10 120 possible states, and a traveling salesperson problem with as little as 82 cities already has more than 10 120 different tours (similar to chess). Biological neurons are slower than the digital switches in computers. An exhaustive search of the problem space exceeds the capacity of current computers for most interesting problems, and it is fairly clear that humans cannot in their lifetime exhaustively search even small fractions of these problem spaces. Yet, humans play chess and solve logistical problems of similar complexity on a daily basis. Even for simple problems humans do not typically engage in exploring even a small fraction of the problem space. This begs the question: How do humans solve problems on a daily basis in a fast and efficient way? Recent work suggests that humans build a problem representation and solve the represented problem-not the problem that is out there. The problem representation that is built and the process used to solve it are constrained by limits of cognitive capacity and a cost-benefit analysis discounting effort and reward. In this article, we argue that better understanding the way humans represent and solve problems using heuristics can help inform how simpler algorithms and representations can be used in artificial intelligence to lower computational complexity, reduce computation time, and facilitate real-time computation in complex problem solving.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Sci-hub)

  • 文章类型: Journal Article
    材料中的缺陷动力学对于从催化到储能系统再到微电子的广泛技术至关重要。材料功能在很大程度上取决于缺陷的性质和组织-它们的排列通常涉及中间或瞬态状态,这些状态为转化提供了很高的障碍。缺乏这些中间状态的知识和这种能量屏障的存在对逆缺陷设计提出了严峻的挑战,尤其是基于梯度的方法。这里,我们提出了一种基于延迟奖励的强化学习(RL)[蒙特卡洛树搜索(MCTS)],可以有效地搜索缺陷配置空间,并使我们能够识别低维材料中的最佳缺陷排列。使用二维MoS2的代表性情况,我们证明了延迟奖励的使用使我们能够有效地对缺陷构型空间进行采样,并克服宽范围缺陷浓度(从1.5%到8%S空位)的能障-系统从初始随机分布的S空位演变为具有与先前实验研究一致的扩展S线缺陷的空位。特征空间中的详细分析使我们能够确定这种缺陷转换和排列的最佳路径。与遗传算法等其他全局优化方案的比较表明,具有延迟奖励的MCTS需要更少的评估,并获得更好的解决方案质量。讨论了各种采样缺陷配置对MoS2中2H至1T相变的影响。总的来说,我们引入了采用延迟奖励的RL策略,可以加速材料缺陷的反向设计,以实现目标功能。
    Defect dynamics in materials are of central importance to a broad range of technologies from catalysis to energy storage systems to microelectronics. Material functionality depends strongly on the nature and organization of defects-their arrangements often involve intermediate or transient states that present a high barrier for transformation. The lack of knowledge of these intermediate states and the presence of this energy barrier presents a serious challenge for inverse defect design, especially for gradient-based approaches. Here, we present a reinforcement learning (RL) [Monte Carlo Tree Search (MCTS)] based on delayed rewards that allow for efficient search of the defect configurational space and allows us to identify optimal defect arrangements in low-dimensional materials. Using a representative case of two-dimensional MoS2, we demonstrate that the use of delayed rewards allows us to efficiently sample the defect configurational space and overcome the energy barrier for a wide range of defect concentrations (from 1.5 to 8% S vacancies)-the system evolves from an initial randomly distributed S vacancies to one with extended S line defects consistent with previous experimental studies. Detailed analysis in the feature space allows us to identify the optimal pathways for this defect transformation and arrangement. Comparison with other global optimization schemes like genetic algorithms suggests that the MCTS with delayed rewards takes fewer evaluations and arrives at a better quality of the solution. The implications of the various sampled defect configurations on the 2H to 1T phase transitions in MoS2 are discussed. Overall, we introduce a RL strategy employing delayed rewards that can accelerate the inverse design of defects in materials for achieving targeted functionality.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Sci-hub)

  • 文章类型: Journal Article
    即使在美国扩大了严重急性呼吸道综合症冠状病毒2(SARS-CoV-2)的疫苗接种,病例将至少在明年在未接种疫苗的人群中徘徊,允许冠状病毒在全国各地的社区继续传播。检测到这些感染,尤其是无症状的,对于在未来几个月阻止病毒的进一步传播至关重要。这将需要积极的监测工作,主动寻找这些未被发现的病例,而不是等待个人到检测地点进行诊断。然而,找到这些无症状病例的口袋(即,热点)类似于大海捞针,因为缺乏流行病学信息来指导决策者分配这些资源,阻碍了在社区内选择何时何地进行测试。利用部分信息进行序贯决策是决策科学中的经典问题,探索v.利用困境。使用类似于用于搜索其他类型的丢失或隐藏对象的方法-土匪算法,从被击落的飞机或地下石油储藏中,我们可以解决探索v.利用权衡面临的积极监测工作,并优化移动测试资源的部署,以最大限度地提高新的SARS-CoV-2诊断的产量。这些强盗算法可以轻松实现,作为SARS-CoV-2主动发现病例的指南。一个简单的Thompson采样算法及其在数据中集成空间相关性的扩展现在被嵌入到一个功能齐全的Web应用程序原型中,允许政策制定者使用这些算法中的任何一个来针对SARS-CoV-2测试。在这种情况下,潜在的测试地点是通过使用UberMedia的移动数据来确定的,目标是哥伦布的高频场所,俄亥俄州,作为该领域算法计划可行性研究的一部分。然而,它很容易适应其他司法管辖区,只需要一组候选测试位置,所有位置之间都有点对点的距离,在选择测试地点时,是否将移动性数据集成到决策中。
    Even as vaccination for severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2) expands in the United States, cases will linger among unvaccinated individuals for at least the next year, allowing the spread of the coronavirus to continue in communities across the country. Detecting these infections, particularly asymptomatic ones, is critical to stemming further transmission of the virus in the months ahead. This will require active surveillance efforts in which these undetected cases are proactively sought out rather than waiting for individuals to present to testing sites for diagnosis. However, finding these pockets of asymptomatic cases (i.e., hotspots) is akin to searching for needles in a haystack as choosing where and when to test within communities is hampered by a lack of epidemiological information to guide decision makers\' allocation of these resources. Making sequential decisions with partial information is a classic problem in decision science, the explore v. exploit dilemma. Using methods-bandit algorithms-similar to those used to search for other kinds of lost or hidden objects, from downed aircraft or underground oil deposits, we can address the explore v. exploit tradeoff facing active surveillance efforts and optimize the deployment of mobile testing resources to maximize the yield of new SARS-CoV-2 diagnoses. These bandit algorithms can be implemented easily as a guide to active case finding for SARS-CoV-2. A simple Thompson sampling algorithm and an extension of it to integrate spatial correlation in the data are now embedded in a fully functional prototype of a web app to allow policymakers to use either of these algorithms to target SARS-CoV-2 testing. In this instance, potential testing locations were identified by using mobility data from UberMedia to target high-frequency venues in Columbus, Ohio, as part of a planned feasibility study of the algorithms in the field. However, it is easily adaptable to other jurisdictions, requiring only a set of candidate test locations with point-to-point distances between all locations, whether or not mobility data are integrated into decision making in choosing places to test.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Sci-hub)

       PDF(Pubmed)

  • 文章类型: Journal Article
    深层生成模型已经显示出设计有效和新颖化学的能力,这可以显著加速生物活性化合物的鉴定。当前的许多型号,然而,使用分子描述符或基于配体的预测方法来引导分子生成朝向期望的性质空间。这将它们的应用限制在数据相对丰富的目标上,忽略那些没有多少数据可以充分训练预测器的地方。此外,基于配体的方法通常将分子生成偏向于先前建立的化学空间,从而限制了他们识别真正新颖化学型的能力。在这项工作中,我们评估了通过Glide-一种基于结构的方法-使用分子对接作为评分函数来指导深度生成模型REINVENT的能力,并将模型性能和行为与基于配体的评分函数进行比较.此外,我们修改了之前发表的MOSES基准测试数据集,以消除对非质子化基团的任何诱导偏倚.我们还提出了一种新的度量数据集多样性的指标,与常用的内部多样性度量相比,重原子计数的分布较少混淆。关于主要发现,我们发现,当优化针对DRD2的对接评分时,该模型将预测的配体亲和力提高到超过已知DRD2活性分子的水平.此外,与基于配体的方法相比,生成的分子占据互补的化学和物理化学空间,和新的物理化学空间相比,已知的DRD2活性分子。此外,基于结构的方法学习生成满足关键残基相互作用的分子,这是仅在考虑蛋白质结构时可用的信息。总的来说,这项工作证明了使用分子对接来指导从头分子生成相对于基于配体的预测因子的优势,新奇,以及识别配体和蛋白质靶标之间关键相互作用的能力。实际上,这种方法在早期命中一代活动中具有应用,以丰富针对特定目标的虚拟库,在以新颖性为中心的项目中,其中从头分子生成要么没有现有的配体知识,要么不应该被它偏颇。
    Deep generative models have shown the ability to devise both valid and novel chemistry, which could significantly accelerate the identification of bioactive compounds. Many current models, however, use molecular descriptors or ligand-based predictive methods to guide molecule generation towards a desirable property space. This restricts their application to relatively data-rich targets, neglecting those where little data is available to sufficiently train a predictor. Moreover, ligand-based approaches often bias molecule generation towards previously established chemical space, thereby limiting their ability to identify truly novel chemotypes. In this work, we assess the ability of using molecular docking via Glide-a structure-based approach-as a scoring function to guide the deep generative model REINVENT and compare model performance and behaviour to a ligand-based scoring function. Additionally, we modify the previously published MOSES benchmarking dataset to remove any induced bias towards non-protonatable groups. We also propose a new metric to measure dataset diversity, which is less confounded by the distribution of heavy atom count than the commonly used internal diversity metric. With respect to the main findings, we found that when optimizing the docking score against DRD2, the model improves predicted ligand affinity beyond that of known DRD2 active molecules. In addition, generated molecules occupy complementary chemical and physicochemical space compared to the ligand-based approach, and novel physicochemical space compared to known DRD2 active molecules. Furthermore, the structure-based approach learns to generate molecules that satisfy crucial residue interactions, which is information only available when taking protein structure into account. Overall, this work demonstrates the advantage of using molecular docking to guide de novo molecule generation over ligand-based predictors with respect to predicted affinity, novelty, and the ability to identify key interactions between ligand and protein target. Practically, this approach has applications in early hit generation campaigns to enrich a virtual library towards a particular target, and also in novelty-focused projects, where de novo molecule generation either has no prior ligand knowledge available or should not be biased by it.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Sci-hub)

       PDF(Pubmed)

公众号