ﻻ يوجد ملخص باللغة العربية
Maximum entropy models provide the least constrained probability distributions that reproduce statistical properties of experimental datasets. In this work we characterize the learning dynamics that maximizes the log-likelihood in the case of large but finite datasets. We first show how the steepest descent dynamics is not optimal as it is slowed down by the inhomogeneous curvature of the model parameters space. We then provide a way for rectifying this space which relies only on dataset properties and does not require large computational efforts. We conclude by solving the long-time limit of the parameters dynamics including the randomness generated by the systematic use of Gibbs sampling. In this stochastic framework, rather than converging to a fixed point, the dynamics reaches a stationary distribution, which for the rectified dynamics reproduces the posterior distribution of the parameters. We sum up all these insights in a rectified Data-Driven algorithm that is fast and by sampling from the parameters posterior avoids both under- and over-fitting along all the directions of the parameters space. Through the learning of pairwise Ising models from the recording of a large population of retina neurons, we show how our algorithm outperforms the steepest descent method.
Many real-world mission-critical applications require continual online learning from noisy data and real-time decision making with a defined confidence level. Probabilistic models and stochastic neural networks can explicitly handle uncertainty in da
We study the typical (median) value of the minimum gap in the quantum version of the Exact Cover problem using Quantum Monte Carlo simulations, in order to understand the complexity of the quantum adiabatic algorithm (QAA) for much larger sizes than
Quantum phase transitions are usually observed in ground states of correlated systems. Remarkably, eigenstate phase transitions can also occur at finite energy density in disordered, isolated quantum systems. Such transitions fall outside the framewo
We have investigated the phase transition in the Heisenberg spin glass using massive numerical simulations to study larger sizes, 48x48x48, than have been attempted before at a spin glass phase transition. A finite-size scaling analysis indicates tha
Image denoising is a well-known and well studied problem, commonly targeting a minimization of the mean squared error (MSE) between the outcome and the original image. Unfortunately, especially for severe noise levels, such Minimum MSE (MMSE) solutio