ترغب بنشر مسار تعليمي؟ اضغط هنا

A Multi-Step Richardson-Romberg Extrapolation Method For Stochastic Approximation

69   0   0.0 ( 0 )
 نشر من قبل Noufel Frikha
 تاريخ النشر 2014
  مجال البحث
والبحث باللغة English
 تأليف Noufel Frikha




اسأل ChatGPT حول البحث

We obtain an expansion of the implicit weak discretization error for the target of stochastic approximation algorithms introduced and studied in [Frikha2013]. This allows us to extend and develop the Richardson-Romberg extrapolation method for Monte Carlo linear estimator (introduced in [Talay & Tubaro 1990] and deeply studied in [Pag{`e}s 2007]) to the framework of stochastic optimization by means of stochastic approximation algorithm. We notably apply the method to the estimation of the quantile of diffusion processes. Numerical results confirm the theoretical analysis and show a significant reduction in the initial computational cost.

قيم البحث

اقرأ أيضاً

In this paper, we consider multi-stage stochastic optimization problems with convex objectives and conic constraints at each stage. We present a new stochastic first-order method, namely the dynamic stochastic approximation (DSA) algorithm, for solvi ng these types of stochastic optimization problems. We show that DSA can achieve an optimal ${cal O}(1/epsilon^4)$ rate of convergence in terms of the total number of required scenarios when applied to a three-stage stochastic optimization problem. We further show that this rate of convergence can be improved to ${cal O}(1/epsilon^2)$ when the objective function is strongly convex. We also discuss variants of DSA for solving more general multi-stage stochastic optimization problems with the number of stages $T > 3$. The developed DSA algorithms only need to go through the scenario tree once in order to compute an $epsilon$-solution of the multi-stage stochastic optimization problem. As a result, the memory required by DSA only grows linearly with respect to the number of stages. To the best of our knowledge, this is the first time that stochastic approximation type methods are generalized for multi-stage stochastic optimization with $T ge 3$.
222 - Liwei Zhang , Yule Zhang , Jia Wu 2019
This paper considers the problem of minimizing a convex expectation function over a closed convex set, coupled with a set of inequality convex expectation constraints. We present a new stochastic approximation type algorithm, namely the stochastic ap proximation proximal method of multipliers (PMMSopt) to solve this convex stochastic optimization problem. We analyze regrets of a stochastic approximation proximal method of multipliers for solving convex stochastic optimization problems. Under mild conditions, we show that this algorithm exhibits ${rm O}(T^{-1/2})$ rate of convergence, in terms of both optimality gap and constraint violation if parameters in the algorithm are properly chosen, when the objective and constraint functions are generally convex, where $T$ denotes the number of iterations. Moreover, we show that, with at least $1-e^{-T^{1/4}}$ probability, the algorithm has no more than ${rm O}(T^{-1/4})$ objective regret and no more than ${rm O}(T^{-1/8})$ constraint violation regret. To the best of our knowledge, this is the first time that such a proximal method for solving expectation constrained stochastic optimization is presented in the literature.
85 - Peng Chen , Qi-Man Shao , Lihu Xu 2020
We view the classical Lindeberg principle in a Markov process setting to establish a universal probability approximation framework by It^{o}s formula and Markov semigroup. As applications, we consider approximating a family of online stochastic gradi ent descents (SGDs) by a stochastic differential equation (SDE) driven by additive Brownian motion, and obtain an approximation error with explicit dependence on the dimension which makes it possible to analyse high dimensional models. We also apply our framework to study stable approximation and normal approximation and obtain their optimal convergence rates (up to a logarithmic correction for normal approximation).
186 - Francis Bach 2020
Richardson extrapolation is a classical technique from numerical analysis that can improve the approximation error of an estimation method by combining linearly several estimates obtained from different values of one of its hyperparameters, without t he need to know in details the inner structure of the original estimation method. The main goal of this paper is to study when Richardson extrapolation can be used within machine learning, beyond the existing applications to step-size adaptations in stochastic gradient descent. We identify two situations where Richardson interpolation can be useful: (1) when the hyperparameter is the number of iterations of an existing iterative optimization algorithm, with applications to averaged gradient descent and Frank-Wolfe algorithms (where we obtain asymptotically rates of $O(1/k^2)$ on polytopes, where $k$ is the number of iterations), and (2) when it is a regularization parameter, with applications to Nesterov smoothing techniques for minimizing non-smooth functions (where we obtain asymptotically rates close to $O(1/k^2)$ for non-smooth functions), and ridge regression. In all these cases, we show that extrapolation techniques come with no significant loss in performance, but with sometimes strong gains, and we provide theoretical justifications based on asymptotic developments for such gains, as well as empirical illustrations on classical problems from machine learning.
Consecutive stochastic 90{deg} polarization switching events, clearly resolved in recent experiments, are described by a new nucleation and growth multi-step model. It extends the classical Kolmogorov-Avrami-Ishibashi approach and includes possible c onsecutive 90{deg}- and parallel 180{deg}-switching events. The model predicts the results of simultaneous time-resolved macroscopic measurements of polarization and strain, performed on a tetragonal Pb(Zr,Ti)O3 ceramic in a wide range of electric fields over a time domain of five orders of the magnitude. It allows the determination of the fractions of individual switching processes, their characteristic switching times, activation fields, and respective Avrami indices.
التعليقات
جاري جلب التعليقات جاري جلب التعليقات
سجل دخول لتتمكن من متابعة معايير البحث التي قمت باختيارها
mircosoft-partner

هل ترغب بارسال اشعارات عن اخر التحديثات في شمرا-اكاديميا