ترغب بنشر مسار تعليمي؟ اضغط هنا

Convergence Rate of the Causal Jacobi Derivative Estimator

137   0   0.0 ( 0 )
 نشر من قبل Dayan Liu
 تاريخ النشر 2011
  مجال البحث
والبحث باللغة English




اسأل ChatGPT حول البحث

Numerical causal derivative estimators from noisy data are essential for real time applications especially for control applications or fluid simulation so as to address the new paradigms in solid modeling and video compression. By using an analytical point of view due to Lanczos cite{C. Lanczos} to this causal case, we revisit $n^{th}$ order derivative estimators originally introduced within an algebraic framework by Mboup, Fliess and Join in cite{num,num0}. Thanks to a given noise level $delta$ and a well-suitable integration length window, we show that the derivative estimator error can be $mathcal{O}(delta ^{frac{q+1}{n+1+q}})$ where $q$ is the order of truncation of the Jacobi polynomial series expansion used. This so obtained bound helps us to choose the values of our parameter estimators. We show the efficiency of our method on some examples.



قيم البحث

اقرأ أيضاً

219 - Taiji Suzuki 2014
In this paper, we investigate the statistical convergence rate of a Bayesian low-rank tensor estimator. Our problem setting is the regression problem where a tensor structure underlying the data is estimated. This problem setting occurs in many pract ical applications, such as collaborative filtering, multi-task learning, and spatio-temporal data analysis. The convergence rate is analyzed in terms of both in-sample and out-of-sample predictive accuracies. It is shown that a near optimal rate is achieved without any strong convexity of the observation. Moreover, we show that the method has adaptivity to the unknown rank of the true tensor, that is, the near optimal rate depending on the true rank is achieved even if it is not known a priori.
The Gaver-Stehfest algorithm is widely used for numerical inversion of Laplace transform. In this paper we provide the first rigorous study of the rate of convergence of the Gaver-Stehfest algorithm. We prove that Gaver-Stehfest approximations conver ge exponentially fast if the target function is analytic in a neighbourhood of a point and they converge at a rate $o(n^{-k})$ if the target function is $(2k+3)$-times differentiable at a point.
This article studies the problem of approximating functions belonging to a Hilbert space $H_d$ with an isotropic or anisotropic Gaussian reproducing kernel, $$ K_d(bx,bt) = expleft(-sum_{ell=1}^dgamma_ell^2(x_ell-t_ell)^2right) mbox{for all} bx,b tinreals^d. $$ The isotropic case corresponds to using the same shape parameters for all coordinates, namely $gamma_ell=gamma>0$ for all $ell$, whereas the anisotropic case corresponds to varying shape parameters $gamma_ell$. We are especially interested in moderate to large $d$.
We propose a novel method to compute a finite difference stencil for Riesz derivative for artibitrary speed of convergence. This method is based on applying a pre-filter to the Grunwald-Letnikov type central difference stencil. The filter is obtained by solving for the inverse of a symmetric Vandemonde matrix and exploiting the relationship between the Taylors series coefficients and fast Fourier transform. The filter costs Oleft(N^{2}right) operations to evaluate for Oleft(h^{N}right) of convergence, where h is the sampling distance. The higher convergence speed should more than offset the overhead with the requirement of the number of nodal points for a desired error tolerance significantly reduced. The benefit of progressive generation of the stencil coefficients for adaptive grid size for dynamic problems with the Grunwald-Letnikov type difference scheme is also kept because of the application of filtering. The higher convergence rate is verified through numerical experiments.
In recent years, contour-based eigensolvers have emerged as a standard approach for the solution of large and sparse eigenvalue problems. Building upon recent performance improvements through non-linear least square optimization of so-called rational filters, we introduce a systematic method to design these filters by minimizing the worst-case convergence ratio and eliminate the parametric dependence on weight functions. Further, we provide an efficient way to deal with the box-constraints which play a central role for the use of iterative linear solvers in contour-based eigensolvers. Indeed, these parameter-free filters consistently minimize the number of iterations and the number of FLOPs to reach convergence in the eigensolver. As a byproduct, our rational filters allow for a simple solution to load balancing when the solution of an interior eigenproblem is approached by the slicing of the sought after spectral interval.
التعليقات
جاري جلب التعليقات جاري جلب التعليقات
سجل دخول لتتمكن من متابعة معايير البحث التي قمت باختيارها
mircosoft-partner

هل ترغب بارسال اشعارات عن اخر التحديثات في شمرا-اكاديميا