ﻻ يوجد ملخص باللغة العربية
A novel and detailed convergence analysis is presented for a greedy algorithm that was previously introduced for operator reconstruction problems in the field of quantum mechanics. This algorithm is based on an offline/online decomposition of the reconstruction process and on an ansatz for the unknown operator obtained by an a priori chosen set of linearly independent matrices. The presented convergence analysis focuses on linear-quadratic (optimization) problems governed by linear differential systems and reveals the strong dependence of the performance of the greedy algorithm on the observability properties of the system and on the ansatz of the basis elements. Moreover, the analysis allows us to use a precise (and in some sense optimal) choice of basis elements for the linear case and led to the introduction of a new and more robust optimized greedy reconstruction algorithm. This optimized approach also applies to nonlinear Hamiltonian reconstruction problems, and its efficiency is demonstrated by numerical experiments.
We suggest a new greedy strategy for convex optimization in Banach spaces and prove its convergent rates under a suitable behavior of the modulus of uniform smoothness of the objective function.
The interest for networks of dynamical systems has been increasing in the past years, especially because of their capability of modeling and describing a large variety of phenomena and behaviors. We propose a technique, based on Wiener filtering, whi
Bandit learning is characterized by the tension between long-term exploration and short-term exploitation. However, as has recently been noted, in settings in which the choices of the learning algorithm correspond to important decisions about individ
Many algorithms for maximizing a monotone submodular function subject to a knapsack constraint rely on the natural greedy heuristic. We present a novel refined analysis of this greedy heuristic which enables us to: $(1)$ reduce the enumeration in the
The alternating direction method of multipliers (ADMM) is a popular method for solving convex separable minimization problems with linear equality constraints. The generalization of the two-block ADMM to the three-block ADMM is not trivial since the