constrained least squares

  • 文章类型: Journal Article
    在本文中,提出了利用约束最小二乘将非线性函数和/或离散数据简化为分段多项式的最优逼近算法。在对时间敏感的应用程序或资源有限的嵌入式系统中,近似函数的运行时间与其准确性一样重要。所提出的算法在确保误差低于指定阈值的同时,以最小的计算成本搜索最佳分段多项式(OPP)。这是通过使用具有最佳顺序和间隔数的平滑分段多项式来实现的。计算成本仅取决于多项式复杂度,即,运行时函数调用的顺序和间隔数。在以往的研究中,用户必须决定一个或所有的订单和间隔的数量。相比之下,OPP近似算法决定了两者。对于最佳逼近,对于特定的目标CPU离线,按升序计算了所有可能的分段多项式组合的计算成本,并将其制成表格。对于给定的样本点,通过约束最小二乘法和随机选择方法对每个组合进行优化。之后,检查近似误差是否低于预定值。如果错误是允许的,该组合被选择为最优近似,或者检查了下一个组合。要验证性能,对几个具有代表性的函数进行了检查和分析。
    In this paper, the optimal approximation algorithm is proposed to simplify non-linear functions and/or discrete data as piecewise polynomials by using the constrained least squares. In time-sensitive applications or in embedded systems with limited resources, the runtime of the approximate function is as crucial as its accuracy. The proposed algorithm searches for the optimal piecewise polynomial (OPP) with the minimum computational cost while ensuring that the error is below a specified threshold. This was accomplished by using smooth piecewise polynomials with optimal order and numbers of intervals. The computational cost only depended on polynomial complexity, i.e., the order and the number of intervals at runtime function call. In previous studies, the user had to decide one or all of the orders and the number of intervals. In contrast, the OPP approximation algorithm determines both of them. For the optimal approximation, computational costs for all the possible combinations of piecewise polynomials were calculated and tabulated in ascending order for the specific target CPU off-line. Each combination was optimized through constrained least squares and the random selection method for the given sample points. Afterward, whether the approximation error was below the predetermined value was examined. When the error was permissible, the combination was selected as the optimal approximation, or the next combination was examined. To verify the performance, several representative functions were examined and analyzed.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    背景:高维转录组分析,无论是通过下一代测序技术还是高通量阵列,可能导致变量分散,数据缺失。数据插补是一种常见的策略,可以通过使用统计技术来填充缺失值,从而最大程度地包含样本。然而,许多数据插补方法繁琐,有引入系统性偏差的风险。
    结果:我们提出了一种新的数据插补方法,使用约束最小二乘法和反问题文献中的算法,并介绍了该技术在miRNA表达分析中的应用。所提出的技术被证明可以更快地提供一个数量级的归算,与文献中的类似方法相比,精度大于或等于精度。
    结论:这项研究为数据填补提供了一种稳健而有效的算法,可以使用,例如,在数据缺失的情况下提高癌症预测的准确性。
    BACKGROUND: High dimensional transcriptome profiling, whether through next generation sequencing techniques or high-throughput arrays, may result in scattered variables with missing data. Data imputation is a common strategy to maximize the inclusion of samples by using statistical techniques to fill in missing values. However, many data imputation methods are cumbersome and risk introduction of systematic bias.
    RESULTS: We present a new data imputation method using constrained least squares and algorithms from the inverse problems literature and present applications for this technique in miRNA expression analysis. The proposed technique is shown to offer an imputation orders of magnitude faster, with greater than or equal accuracy when compared to similar methods from the literature.
    CONCLUSIONS: This study offers a robust and efficient algorithm for data imputation, which can be used, e.g., to improve cancer prediction accuracy in the presence of missing data.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    Navigation accuracy is one of the key performance indicators of an inertial navigation system (INS). Requirements for an accuracy assessment of an INS in a real work environment are exceedingly urgent because of enormous differences between real work and laboratory test environments. An attitude accuracy assessment of an INS based on the intensified high dynamic star tracker (IHDST) is particularly suitable for a real complex dynamic environment. However, the coupled systematic coordinate errors of an INS and the IHDST severely decrease the attitude assessment accuracy of an INS. Given that, a high-accuracy decoupling estimation method of the above systematic coordinate errors based on the constrained least squares (CLS) method is proposed in this paper. The reference frame of the IHDST is firstly converted to be consistent with that of the INS because their reference frames are completely different. Thereafter, the decoupling estimation model of the systematic coordinate errors is established and the CLS-based optimization method is utilized to estimate errors accurately. After compensating for error, the attitude accuracy of an INS can be assessed based on IHDST accurately. Both simulated experiments and real flight experiments of aircraft are conducted, and the experimental results demonstrate that the proposed method is effective and shows excellent performance for the attitude accuracy assessment of an INS in a real work environment.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Sci-hub)

  • 文章类型: Journal Article
    生物分子筛选研究经常寻找最有可能使生物化学或基于细胞的测定系统产生强烈连续响应的化合物。每种化合物都测试了几种剂量,并且假定,如果存在剂量反应关系,这种关系遵循单调曲线,通常是中值效应方程的一个版本。然而,没有关系的零假设不能用这个方程进行统计检验。我们使用这个方程的线性化版本来定义药理作用大小的度量,并使用此措施将所研究的化合物按其产生强烈反应的总体能力进行排序。用此方法可以测试零假设,即没有一个特定化合物的检查剂量产生强烈的反应。所提出的办法是基于新的统计模子的重要概念反响检测限,在连续反应的剂量反应数据分析中通常被忽略的概念。该方法通过一项研究的数据进行了说明,该研究寻找了中和人类免疫缺陷病毒对脑胶质母细胞瘤细胞的感染的化合物。
    Biomolecular screening research frequently searches for the chemical compounds that are most likely to make a biochemical or cell-based assay system produce a strong continuous response. Several doses are tested with each compound and it is assumed that, if there is a dose-response relationship, the relationship follows a monotonic curve, usually a version of the median-effect equation. However, the null hypothesis of no relationship cannot be statistically tested using this equation. We used a linearized version of this equation to define a measure of pharmacological effect size, and use this measure to rank the investigated compounds in order of their overall capability to produce strong responses. The null hypothesis that none of the examined doses of a particular compound produced a strong response can be tested with this approach. The proposed approach is based on a new statistical model of the important concept of response detection limit, a concept that is usually neglected in the analysis of dose-response data with continuous responses. The methodology is illustrated with data from a study searching for compounds that neutralize the infection by a human immunodeficiency virus of brain glioblastoma cells.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    The use of ROC curves in evaluating a continuous or ordinal biomarker for the discrimination of two populations is commonplace. However, in many settings, marker measurements above or below a certain value cannot be obtained. In this paper, we study the construction of a smooth ROC curve (or surface in the case of three populations) when there is a lower or upper limit of detection. We propose the use of spline models that incorporate monotonicity constraints for the cumulative hazard function of the marker distribution. The proposed technique is computationally stable and simulation results showed a satisfactory performance. Other observed covariates can be also accommodated by this spline-based approach.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Sci-hub)

公众号