Hyperparameter

超参数
  • 文章类型: Journal Article
    肌肉减少性肥胖(SO)的特征是伴随的肌肉减少症和肥胖,并存在残疾的高风险,发病率,和老年人的死亡率。然而,缺乏基于序贯神经网络SO研究的预测以及体能因子与SO之间的关系。本研究旨在通过关注身体健康因素来开发老年人SO的预测模型。使用顺序神经网络分析了参加全民健身计划的韩国老年人的综合数据集。使用人体测量方程将阑尾骨骼肌/体重定义为SO。独立变量包括身体脂肪(BF,%),腰围,收缩压和舒张压,和各种体能因素。因变量是二元结果(可能的SO与正常)。我们分析了超参数调整和分层K倍验证,以优化预测模型。女性的SO患病率(13.81%)明显高于男性,突出性别差异。优化后的神经网络模型和Shapley加法解释分析表明,验证准确率高达93.1%,BF%和绝对握力成为SO最有影响力的预测因子。这项研究为老年人的SO提供了一个高度准确的预测模型,强调BF%和绝对握力的关键作用。我们确认了BF,绝对握力,作为关键的SO预测因子。我们的发现强调了SO的性别特异性以及身体健康因素在预测中的重要性。
    Sarcopenic obesity (SO) is characterized by concomitant sarcopenia and obesity and presents a high risk of disability, morbidity, and mortality among older adults. However, predictions based on sequential neural network SO studies and the relationship between physical fitness factors and SO are lacking. This study aimed to develop a predictive model for SO in older adults by focusing on physical fitness factors. A comprehensive dataset of older Korean adults participating in national fitness programs was analyzed using sequential neural networks. Appendicular skeletal muscle/body weight was defined as SO using an anthropometric equation. Independent variables included body fat (BF, %), waist circumference, systolic and diastolic blood pressure, and various physical fitness factors. The dependent variable was a binary outcome (possible SO vs normal). We analyzed hyperparameter tuning and stratified K-fold validation to optimize a predictive model. The prevalence of SO was significantly higher in women (13.81%) than in men, highlighting sex-specific differences. The optimized neural network model and Shapley Additive Explanations analysis demonstrated a high validation accuracy of 93.1%, with BF% and absolute grip strength emerging as the most influential predictors of SO. This study presents a highly accurate predictive model for SO in older adults, emphasizing the critical roles of BF% and absolute grip strength. We identified BF, absolute grip strength, and sit-and-reach as key SO predictors. Our findings underscore the sex-specific nature of SO and the importance of physical fitness factors in its prediction.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    深度强化学习(DRL)在不同的领域和应用中获得了广泛的采用,主要是由于其在具有高维状态和动作的空间中解决复杂的决策问题的能力。深度确定性策略梯度(DDPG)是一种众所周知的DRL算法,采用演员-批评方法,综合基于价值和基于策略的强化学习方法的优势。这项研究的目的是全面研究最新发展,模式,障碍,以及与DDPG相关的潜在机会。使用相关的学术数据库进行了系统的搜索(Scopus,WebofScience,和ScienceDirect)确定过去五年(2018-2023年)发表的85项相关研究。我们全面概述了DDPG的关键概念和组件,包括它的配方,实施,和训练。然后,我们重点介绍了DDPG的各种应用和领域,包括自动驾驶,无人机,资源分配,通信和物联网,机器人,和金融。此外,我们提供了DDPG与其他DRL算法和传统RL方法的深入比较,突出它的优点和缺点。我们相信,这次审查将是研究人员的重要资源,为他们提供有关DRL和DDPG领域使用的方法和技术的宝贵见解。
    Deep Reinforcement Learning (DRL) has gained significant adoption in diverse fields and applications, mainly due to its proficiency in resolving complicated decision-making problems in spaces with high-dimensional states and actions. Deep Deterministic Policy Gradient (DDPG) is a well-known DRL algorithm that adopts an actor-critic approach, synthesizing the advantages of value-based and policy-based reinforcement learning methods. The aim of this study is to provide a thorough examination of the latest developments, patterns, obstacles, and potential opportunities related to DDPG. A systematic search was conducted using relevant academic databases (Scopus, Web of Science, and ScienceDirect) to identify 85 relevant studies published in the last five years (2018-2023). We provide a comprehensive overview of the key concepts and components of DDPG, including its formulation, implementation, and training. Then, we highlight the various applications and domains of DDPG, including Autonomous Driving, Unmanned Aerial Vehicles, Resource Allocation, Communications and the Internet of Things, Robotics, and Finance. Additionally, we provide an in-depth comparison of DDPG with other DRL algorithms and traditional RL methods, highlighting its strengths and weaknesses. We believe that this review will be an essential resource for researchers, offering them valuable insights into the methods and techniques utilized in the field of DRL and DDPG.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    单细胞分辨系统生物学方法,包括基于物质和成像的测量模式,生成大量表征细胞群体异质性的高维数据。表示学习方法通常用于分析这些复杂的,通过将高维数据投影到低维嵌入中。这有助于结构的解释和询问,动力学,和细胞异质性的调节。反映了它们在分析不同单细胞数据类型中的核心作用,存在无数的表征学习方法,新方法不断涌现。这里,我们对比了跨越统计的表示学习方法的一般特征,流形学习,和神经网络方法。我们考虑使用单细胞数据表示学习中涉及的关键步骤,包括数据预处理,超参数优化,下游分析,和生物验证。还强调了将这些步骤联系起来的相互依存关系和突发事件。此概述旨在指导研究人员进行选择,应用程序,以及当前和未来单细胞研究应用的表征学习策略的优化。
    Single-cell-resolved systems biology methods, including omics- and imaging-based measurement modalities, generate a wealth of high-dimensional data characterizing the heterogeneity of cell populations. Representation learning methods are routinely used to analyze these complex, high-dimensional data by projecting them into lower-dimensional embeddings. This facilitates the interpretation and interrogation of the structures, dynamics, and regulation of cell heterogeneity. Reflecting their central role in analyzing diverse single-cell data types, a myriad of representation learning methods exist, with new approaches continually emerging. Here, we contrast general features of representation learning methods spanning statistical, manifold learning, and neural network approaches. We consider key steps involved in representation learning with single-cell data, including data pre-processing, hyperparameter optimization, downstream analysis, and biological validation. Interdependencies and contingencies linking these steps are also highlighted. This overview is intended to guide researchers in the selection, application, and optimization of representation learning strategies for current and future single-cell research applications.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    Objective.本文旨在提出一种先进的方法,用于使用具有计算机断层扫描(CT)图像的自动技术评估肺结节,以在早期检测肺癌。方法。所提出的方法在卷积神经网络(CNN)中利用固定大小的3×3内核进行相关特征提取。网络体系结构包括13层,包括六个卷积层,用于深度局部和全局特征提取。通过合并基于迁移学习的EfficientNetV_2网络(TLEV2N)来增强结节检测体系结构,以提高训练性能。结节的分类是通过整合CNN的EfficientNet_V2架构来实现的,以实现更准确的良性和恶性分类。网络架构经过微调,可以使用深度网络提取相关特征,同时通过合适的超参数保持性能。主要结果。该方法显著降低了假阴性率,该网络的准确率为97.56%,特异性为98.4%。使用3×3内核提供了对微小像素变化的有价值的见解,并能够在更广泛的形态学水平上提取信息。网络对微调初始值的连续响应允许进一步优化的可能性,导致能够评估多种胸部CT数据集的标准化系统的设计。意义。本文通过分析低剂量CT图像,重点介绍了非侵入性技术在早期发现肺癌方面的潜力。所提出的方法提高了检测肺结节的准确性,并有可能提高早期肺癌检测的整体性能。通过重新配置所提出的方法,可以进一步改进以优化结果,并有助于开发用于评估不同胸部CT数据集的标准化系统.
    Objective. This paper aims to propose an advanced methodology for assessing lung nodules using automated techniques with computed tomography (CT) images to detect lung cancer at an early stage.Approach. The proposed methodology utilizes a fixed-size 3 × 3 kernel in a convolution neural network (CNN) for relevant feature extraction. The network architecture comprises 13 layers, including six convolution layers for deep local and global feature extraction. The nodule detection architecture is enhanced by incorporating a transfer learning-based EfficientNetV_2 network (TLEV2N) to improve training performance. The classification of nodules is achieved by integrating the EfficientNet_V2 architecture of CNN for more accurate benign and malignant classification. The network architecture is fine-tuned to extract relevant features using a deep network while maintaining performance through suitable hyperparameters.Main results. The proposed method significantly reduces the false-negative rate, with the network achieving an accuracy of 97.56% and a specificity of 98.4%. Using the 3 × 3 kernel provides valuable insights into minute pixel variation and enables the extraction of information at a broader morphological level. The continuous responsiveness of the network to fine-tune initial values allows for further optimization possibilities, leading to the design of a standardized system capable of assessing diversified thoracic CT datasets.Significance. This paper highlights the potential of non-invasive techniques for the early detection of lung cancer through the analysis of low-dose CT images. The proposed methodology offers improved accuracy in detecting lung nodules and has the potential to enhance the overall performance of early lung cancer detection. By reconfiguring the proposed method, further advancements can be made to optimize outcomes and contribute to developing a standardized system for assessing diverse thoracic CT datasets.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    统计欺诈检测包括建立一个系统,自动选择所有案件的子集(保险索赔,金融交易,等。),这是最有趣的进一步调查。需要这种系统的原因是案件总数通常比实际可以手动调查的案件总数高得多,并且欺诈往往非常罕见。Further,研究者通常仅限于控制有限数量的k个病例,由于资源有限。分配这些资源的最有效方式是尝试选择具有最高欺诈概率的k个案例。通常必须对用于此目的的预测模型进行正则化,以避免过度拟合并因此避免不良的预测性能。损失函数,表示欺诈损失,提出了通过调整参数选择模型复杂度的方法。进行模拟研究以找到用于验证的最佳设置。Further,将拟议程序的性能与最相关的竞争程序进行比较,根据接受者工作特征曲线下面积(AUC),在一组模拟中,以及信用卡默认数据集。与根据AUC选择模型相比,通过欺诈损失选择模型的复杂性在欺诈损失方面产生了可比或更好的结果。
    Statistical fraud detection consists in making a system that automatically selects a subset of all cases (insurance claims, financial transactions, etc.) that are the most interesting for further investigation. The reason why such a system is needed is that the total number of cases typically is much higher than one realistically could investigate manually and that fraud tends to be quite rare. Further, the investigator is typically limited to controlling a restricted number k of cases, due to limited resources. The most efficient manner of allocating these resources is then to try selecting the k cases with the highest probability of being fraudulent. The prediction model used for this purpose must normally be regularised to avoid overfitting and consequently bad prediction performance. A loss function, denoted the fraud loss, is proposed for selecting the model complexity via a tuning parameter. A simulation study is performed to find the optimal settings for validation. Further, the performance of the proposed procedure is compared to the most relevant competing procedure, based on the area under the receiver operating characteristic curve (AUC), in a set of simulations, as well as on a credit card default dataset. Choosing the complexity of the model by the fraud loss resulted in either comparable or better results in terms of the fraud loss than choosing it according to the AUC.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    利用深度学习网络进行语义分割已成为从高分辨率遥感图像中提取目标的重要方法。与传统的卷积神经网络(CNN)相比,VisionTransformer网络在语义分割方面的性能显着提高。视觉转换器网络具有与CNN不同的架构。映像修补程序,线性嵌入,多头自我注意力(MHSA)是几个主要的超参数。我们应该如何配置它们以提取VHR图像中的对象以及它们如何影响网络的准确性是尚未得到充分研究的主题。本文探讨了视觉变压器网络在从极高分辨率(VHR)图像中提取建筑物足迹中的作用。设计并比较了具有不同超参数值的变压器模型,并分析了它们对准确性的影响。结果表明,较小的图像块和较高维嵌入可以获得更好的精度。此外,基于Transformer的网络被证明是可扩展的,并且可以使用具有与卷积神经网络相当的模型大小和训练时间的通用规模图形处理单元(GPU)进行训练,同时实现更高的精度。该研究为视觉变压器网络在使用VHR图像进行对象提取中的潜力提供了有价值的见解。
    Semantic segmentation with deep learning networks has become an important approach to the extraction of objects from very high-resolution remote sensing images. Vision Transformer networks have shown significant improvements in performance compared to traditional convolutional neural networks (CNNs) in semantic segmentation. Vision Transformer networks have different architectures to CNNs. Image patches, linear embedding, and multi-head self-attention (MHSA) are several of the main hyperparameters. How we should configure them for the extraction of objects in VHR images and how they affect the accuracy of networks are topics that have not been sufficiently investigated. This article explores the role of vision Transformer networks in the extraction of building footprints from very-high-resolution (VHR) images. Transformer-based models with different hyperparameter values were designed and compared, and their impact on accuracy was analyzed. The results show that smaller image patches and higher-dimension embeddings result in better accuracy. In addition, the Transformer-based network is shown to be scalable and can be trained with general-scale graphics processing units (GPUs) with comparable model sizes and training times to convolutional neural networks while achieving higher accuracy. The study provides valuable insights into the potential of vision Transformer networks in object extraction using VHR images.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    COVID-19,不断发展和提出越来越重要的问题,影响了人类健康,造成了无数人死亡。它是一种发病率和死亡率都很高的传染病。这种疾病的传播也是对人类健康的重大威胁,尤其是在发展中国家。这项研究提出了一种称为基于混洗牧羊人优化的广义深度卷积模糊网络(SSO-GDCFN)的方法来诊断COVID-19疾病状态,类型,和恢复的类别。结果表明,该方法的准确率高达99.99%;同样,精确度为99.98%;灵敏度/召回率为100%;特异性为95%;κ为0.965%;AUC为0.88%;MSE小于0.07%以及25秒。通过将所提出的方法的仿真结果与几种传统技术的仿真结果进行比较,证实了所提出方法的性能。实验结果表明,与传统方法相比,COVID-19阶段的分类具有很强的性能和很高的准确性,并且重分类最少。
    COVID-19, continually developing and raising increasingly significant issues, has impacted human health and caused countless deaths. It is an infectious disease with a high incidence and mortality rate. The spread of the disease is also a significant threat to human health, especially in the developing world. This study suggests a method called shuffle shepherd optimization-based generalized deep convolutional fuzzy network (SSO-GDCFN) to diagnose the COVID-19 disease state, types, and recovered categories. The results show that the accuracy of the proposed method is as high as 99.99%; similarly, precision is 99.98%; sensitivity/recall is 100%; specificity is 95%; kappa is 0.965%; AUC is 0.88%; and MSE is less than 0.07% as well as 25 s. Moreover, the performance of the suggested method has been confirmed by comparison of the simulation results from the proposed approach with those from several traditional techniques. The experimental findings demonstrate strong performance and high accuracy for categorizing COVID-19 stages with minimal reclassifications over the conventional methods.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    童年的逆境会对以后的生活产生广泛而持久的影响。但是,造成这些影响的机制是什么?本文汇集了有关探索-利用权衡的认知科学文献,关于早期逆境的实证文献,以及进化生物学中关于“生活史”的文献来解释早期经验如何影响以后的生活。我们提出了一种潜在的机制:早期经验会影响“超参数”,从而决定了勘探与开发之间的平衡。逆境可能会加速从探索到开发的转变,对成人的大脑和思想有广泛而持久的影响。这些影响可能是由生活史适应产生的,这些适应利用早期经验来调整发展和学习,以适应生物体及其环境的可能的未来状态。
    Childhood adversity can have wide-ranging and long-lasting effects on later life. But what are the mechanisms that are responsible for these effects? This article brings together the cognitive science literature on explore-exploit tradeoffs, the empirical literature on early adversity, and the literature in evolutionary biology on \'life history\' to explain how early experience influences later life. We propose one potential mechanism: early experiences influence \'hyperparameters\' that determine the balance between exploration and exploitation. Adversity might accelerate a shift from exploration to exploitation, with broad and enduring effects on the adult brain and mind. These effects may be produced by life-history adaptations that use early experience to tailor development and learning to the likely future states of an organism and its environment.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    基于贝叶斯定律的层析成像图像重建的大多数惩罚最大似然方法包括可自由调整的超参数,以平衡数据保真度项和特定噪声分辨率权衡的先验/惩罚项。在许多应用中,超参数是通过反复试验的方式根据经验确定的,然后从多次迭代重建中选择最优结果。这些惩罚方法不仅因其迭代性质而耗时,还需要手动调整。本研究旨在研究一种基于理论的贝叶斯图像重建策略,无需可自由调整的超参数,大大节省时间和计算资源。贝叶斯图像重建问题由两个概率密度函数(PDF)表示,一个用于数据保真度项,另一个用于先前项。在制定这些PDF时,我们引入两个参数。虽然这两个参数确保PDF完全描述了数据和先前的术语,它们不能由获取的数据确定;因此,它们被称为完全但不可观察的参数。估计这两个参数成为可能的条件期望和最大化的图像重建,给定获得的数据和PDF。这导致了一个迭代算法,联合估计两个参数,并通过最大化后验概率来计算待重建图像,表示为联合参数贝叶斯。除了理论上的表述,进行了综合仿真实验,分析了迭代联合参数贝叶斯方法的停止准则。最后,鉴于数据,通过满足数据似然性和先验概率的PDF条件,在没有任何可自由调整的超参数的情况下获得最佳重建,并满足停止标准。此外,通过初始化等因素研究了联合参数贝叶斯的稳定性,PDF规范,并以迭代方式重新归一化。体模模拟和临床患者数据结果表明,与传统方法相比,联合参数贝叶斯可以提供相当的重建图像质量。但重建时间要少得多。要查看算法对不同类型噪声的响应,在仿真数据中引入了三种常见的噪声模型,包括高斯白噪声到对数正弦图数据,泊松信号相关噪声对对数后正弦图数据和泊松噪声对对数前传输数据。高斯白噪声的实验结果表明,联合参数贝叶斯方法估计的两个参数与仿真吻合良好。观察到,为满足先前的PDF而引入的参数对停止所有三个噪声模型的迭代过程更敏感。稳定性研究表明,通过滤波反投影获得的初始图像非常鲁棒。临床患者数据证明了所提出的联合参数贝叶斯和停止标准的有效性。
    Most penalized maximum likelihood methods for tomographic image reconstruction based on Bayes\' law include a freely adjustable hyperparameter to balance the data fidelity term and the prior/penalty term for a specific noise-resolution tradeoff. The hyperparameter is determined empirically via a trial-and-error fashion in many applications, which then selects the optimal result from multiple iterative reconstructions. These penalized methods are not only time-consuming by their iterative nature, but also require manual adjustment. This study aims to investigate a theory-based strategy for Bayesian image reconstruction without a freely adjustable hyperparameter, to substantially save time and computational resources. The Bayesian image reconstruction problem is formulated by two probability density functions (PDFs), one for the data fidelity term and the other for the prior term. When formulating these PDFs, we introduce two parameters. While these two parameters ensure the PDFs completely describe the data and prior terms, they cannot be determined by the acquired data; thus, they are called complete but unobservable parameters. Estimating these two parameters becomes possible under the conditional expectation and maximization for the image reconstruction, given the acquired data and the PDFs. This leads to an iterative algorithm, which jointly estimates the two parameters and computes the to-be reconstructed image by maximizing a posteriori probability, denoted as joint-parameter-Bayes. In addition to the theoretical formulation, comprehensive simulation experiments are performed to analyze the stopping criterion of the iterative joint-parameter-Bayes method. Finally, given the data, an optimal reconstruction is obtained without any freely adjustable hyperparameter by satisfying the PDF condition for both the data likelihood and the prior probability, and by satisfying the stopping criterion. Moreover, the stability of joint-parameter-Bayes is investigated through factors such as initialization, the PDF specification, and renormalization in an iterative manner. Both phantom simulation and clinical patient data results show that joint-parameter-Bayes can provide comparable reconstructed image quality compared to the conventional methods, but with much less reconstruction time. To see the response of the algorithm to different types of noise, three common noise models are introduced to the simulation data, including white Gaussian noise to post-log sinogram data, Poisson-like signal-dependent noise to post-log sinogram data and Poisson noise to the pre-log transmission data. The experimental outcomes of the white Gaussian noise reveal that the two parameters estimated by the joint-parameter-Bayes method agree well with simulations. It is observed that the parameter introduced to satisfy the prior\'s PDF is more sensitive to stopping the iteration process for all three noise models. A stability investigation showed that the initial image by filtered back projection is very robust. Clinical patient data demonstrated the effectiveness of the proposed joint-parameter-Bayes and stopping criterion.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    背景:在外束放射治疗中,需要一个预测模型来补偿影响辐射剂量输送准确性的时间系统延迟。本研究集中于对七个深度人工神经网络的全面比较,以提出准确可靠的预测模型。
    方法:用800个呼吸信号对7个深度预测模型进行了训练和测试。在这方面,开发了一种非序列相关的超参数优化算法,以找到所有模型的最佳参数配置。均方根误差(RMSE),平均绝对误差,归一化RMSE,统计F检验也用于评估网络性能。
    结果:总体而言,与以前的研究相比,调整超参数可以使所有模型提高25%-30%。所有模型之间的比较还表明,RMSE=0.108±0.068mm的门控递归单位(GRU)以更高的精度和更好的性能预测呼吸信号。
    结论:总体而言,调整GRU模型中的超参数显示出比CyberKnifeVSI系统中使用的混合预测模型更好的结果,以补偿115毫秒的系统延迟。此外,结果表明,调整后的参数对每个模型的预测精度都有显著影响。
    BACKGROUND: In external beam radiotherapy, a prediction model is required to compensate for the temporal system latency that affects the accuracy of radiation dose delivery. This study focused on a thorough comparison of seven deep artificial neural networks to propose an accurate and reliable prediction model.
    METHODS: Seven deep predictor models are trained and tested with 800 breathing signals. In this regard, a nonsequential-correlated hyperparameter optimization algorithm is developed to find the best configuration of parameters for all models. The root mean square error (RMSE), mean absolute error, normalized RMSE, and statistical F-test are also used to evaluate network performance.
    RESULTS: Overall, tuning the hyperparameters results in a 25%-30% improvement for all models compared to previous studies. The comparison between all models also shows that the gated recurrent unit (GRU) with RMSE = 0.108 ± 0.068 mm predicts respiratory signals with higher accuracy and better performance.
    CONCLUSIONS: Overall, tuning the hyperparameters in the GRU model demonstrates a better result than the hybrid predictor model used in the CyberKnife VSI system to compensate for the 115 ms system latency. Additionally, it is demonstrated that the tuned parameters have a significant impact on the prediction accuracy of each model.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

公众号