No Arabic abstract
We apply the power-of-two-choices paradigm to a random walk on a graph: rather than moving to a uniform random neighbour at each step, a controller is allowed to choose from two independent uniform random neighbours. We prove that this allows the controller to significantly accelerate the hitting and cover times in several natural graph classes. In particular, we show that the cover time becomes linear in the number $n$ of vertices on discrete tori and bounded degree trees, of order $mathcal{O}(n log log n)$ on bounded degree expanders, and of order $mathcal{O}(n (log log n)^2)$ on the ErdH{o}s-R{e}nyi random graph in a certain sparsely connected regime. We also consider the algorithmic question of computing an optimal strategy, and prove a dichotomy in efficiency between computing strategies for hitting and cover times.
Random factor graphs provide a powerful framework for the study of inference problems such as decoding problems or the stochastic block model. Information-theoretically the key quantity of interest is the mutual information between the observed factor graph and the underlying ground truth around which the factor graph was created; in the stochastic block model, this would be the planted partition. The mutual information gauges whether and how well the ground truth can be inferred from the observable data. For a very general model of random factor graphs we verify a formula for the mutual information predicted by physics techniques. As an application we prove a conjecture about low-density generator matrix codes from [Montanari: IEEE Transactions on Information Theory 2005]. Further applications include phase transitions of the stochastic block model and the mixed $k$-spin model from physics.
We consider the proportion of generalized visible lattice points in the plane visited by random walkers. Our work concerns the visible lattice points in random walks in three aspects: (1) generalized visibility along curves; (2) one random walker visible from multiple watchpoints; (3) simultaneous visibility of multiple random walkers. Moreover, we found new phenomenon in the case of multiple random walkers: for visibility along a large class of curves and for any number of random walkers, the proportion of steps at which all random walkers are visible simultaneously is almost surely larger than a positive constant.
This is the second in a series of articles devoted to showing that a typical covering map of large degree to a fixed, regular graph has its new adjacency eigenvalues within the bound conjectured by Alon for random regular graphs. The first main result in this article concerns the function $f(k,n)$ defined as the number of SNBC (strictly non-backtracking closed) walks of length $k$ of a given homotopy type in a random covering graph of degree $n$ of a fixed graph. We prove the existence of asymptotic expansions in powers of $1/n$ for $f(k,n)$, where the coefficients---functions of $k$---are proven to have some desirable properties; namely, these coefficients are approximately a sum of polynomials times exponential functions. The second main result is a generalization of the first, where the number of SNBC walks of length $k$ is multiplied by an indicator function that the covering graph contains a certain type of {em tangle}; the second result requires more terminology, although its proof uses the same basic tools used to prove the first result. % The motivation for the second main result will be clear in % the third article in this series of articles. The results in this article are mostly straightforward generalizations of methods used in previous works. However, this article (1) factors these methods into a number of short, conceptually simple, and independent parts, (2) writes each independent part in more general terms, and (3) significantly simplifies of one of the previous computations. As such we expect that this article will make it easier to apply trace methods to related models of random graphs.
We study random walks on the giant component of the ErdH{o}s-Renyi random graph ${cal G}(n,p)$ where $p=lambda/n$ for $lambda>1$ fixed. The mixing time from a worst starting point was shown by Fountoulakis and Reed, and independently by Benjamini, Kozma and Wormald, to have order $log^2 n$. We prove that starting from a uniform vertex (equivalently, from a fixed vertex conditioned to belong to the giant) both accelerates mixing to $O(log n)$ and concentrates it (the cutoff phenomenon occurs): the typical mixing is at $( u {bf d})^{-1}log n pm (log n)^{1/2+o(1)}$, where $ u$ and ${bf d}$ are the speed of random walk and dimension of harmonic measure on a ${rm Poisson}(lambda)$-Galton-Watson tree. Analogous results are given for graphs with prescribed degree sequences, where cutoff is shown both for the simple and for the non-backtracking random walk.
We develop a framework for the rigorous analysis of focused stochastic local search algorithms. These are algorithms that search a state space by repeatedly selecting some constraint that is violated in the current state and moving to a random nearby state that addresses the violation, while hopefully not introducing many new ones. An important class of focused local search algorithms with provable performance guarantees has recently arisen from algorithmizations of the Lov{a}sz Local Lemma (LLL), a non-constructive tool for proving the existence of satisfying states by introducing a background measure on the state space. While powerful, the state transitions of algorithms in this class must be, in a precise sense, perfectly compatible with the background measure. In many applications this is a very restrictive requirement and one needs to step outside the class. Here we introduce the notion of emph{measure distortion} and develop a framework for analyzing arbitrary focused stochastic local search algorithms, recovering LLL algorithmizations as the special case of no distortion. Our framework takes as input an arbitrary such algorithm and an arbitrary probability measure and shows how to use the measure as a yardstick of algorithmic progress, even for algorithms designed independently of the measure.