Improved Convergence Rates for the Orthogonal Greedy Algorithm


Abstract in English

We analyze the orthogonal greedy algorithm when applied to dictionaries $mathbb{D}$ whose convex hull has small entropy. We show that if the metric entropy of the convex hull of $mathbb{D}$ decays at a rate of $O(n^{-frac{1}{2}-alpha})$ for $alpha > 0$, then the orthogonal greedy algorithm converges at the same rate. This improves upon the well-known $O(n^{-frac{1}{2}})$ convergence rate of the orthogonal greedy algorithm in many cases, most notably for dictionaries corresponding to shallow neural networks. Finally, we show that these improved rates are sharp under the given entropy decay assumptions.

Download