No Arabic abstract
In order to stabilize the behavior of noisy systems, confining it around a desirable state, an effort is required to suppress the intrinsic noise. This noise suppression task entails a cost. For the important case of thermal noise in an overdamped system, we show that the minimum cost is achieved when the system control parameters are held constant: any additional deterministic or random modulation produces an increase of the cost. We discuss the implications of this phenomenon for those overdamped systems whose control parameters are intrinsically noisy, presenting a case study based on the example of a Brownian particle optically trapped in an oscillating potential.
We consider the effect of introducing a small number of non-aligning agents in a well-formed flock. To this end, we modify a minimal model of active Brownian particles with purely repulsive (excluded volume) forces to introduce an alignment interaction that will be experienced by all the particles except for a small minority of dissenters. We find that even a very small fraction of dissenters disrupts the flocking state. Strikingly, these motile dissenters are much more effective than an equal number of static obstacles in breaking up the flock. For the studied system sizes we obtain clear evidence of scale invariance at the flocking-disorder transition point and the system can be effectively described with a finite-size scaling formalism. We develop a continuum model for the system which reveals that dissenters act like annealed noise on aligners, with a noise strength that grows with the persistence of the dissenters dynamics.
Data science and machine learning (DS/ML) are at the heart of the recent advancements of many Artificial Intelligence (AI) applications. There is an active research thread in AI, autoai, that aims to develop systems for automating end-to-end the DS/ML Lifecycle. However, do DS and ML workers really want to automate their DS/ML workflow? To answer this question, we first synthesize a human-centered AutoML framework with 6 User Role/Personas, 10 Stages and 43 Sub-Tasks, 5 Levels of Automation, and 5 Types of Explanation, through reviewing research literature and marketing reports. Secondly, we use the framework to guide the design of an online survey study with 217 DS/ML workers who had varying degrees of experience, and different user roles matching to our 6 roles/personas. We found that different user personas participated in distinct stages of the lifecycle -- but not all stages. Their desired levels of automation and types of explanation for AutoML also varied significantly depending on the DS/ML stage and the user persona. Based on the survey results, we argue there is no rationale from user needs for complete automation of the end-to-end DS/ML lifecycle. We propose new next steps for user-controlled DS/ML automation.
There is a longstanding discrepancy between the observed Galactic classical nova rate of $sim 10$ yr$^{-1}$ and the predicted rate from Galactic models of $sim 30$--50 yr$^{-1}$. One explanation for this discrepancy is that many novae are hidden by interstellar extinction, but the degree to which dust can obscure novae is poorly constrained. We use newly available all-sky three-dimensional dust maps to compare the brightness and spatial distribution of known novae to that predicted from relatively simple models in which novae trace Galactic stellar mass. We find that only half ($sim 48$%) of novae are expected to be easily detectable ($g lesssim 15$) with current all-sky optical surveys such as the All-Sky Automated Survey for Supernovae (ASAS-SN). This fraction is much lower than previously estimated, showing that dust does substantially affect nova detection in the optical. By comparing complementary survey results from ASAS-SN, OGLE-IV, and the Palomar Gattini IR-survey in the context of our modeling, we find a tentative Galactic nova rate of $sim 40$ yr$^{-1}$, though this could decrease to as low as $sim 30$ yr$^{-1}$ depending on the assumed distribution of novae within the Galaxy. These preliminary estimates will be improved in future work through more sophisticated modeling of nova detection in ASAS-SN and other surveys.
Locally checkable labeling problems (LCLs) are distributed graph problems in which a solution is globally feasible if it is locally feasible in all constant-radius neighborhoods. Vertex colorings, maximal independent sets, and maximal matchings are examples of LCLs. On the one hand, it is known that some LCLs benefit exponentially from randomness---for example, any deterministic distributed algorithm that finds a sinkless orientation requires $Theta(log n)$ rounds in the LOCAL model, while the randomized complexity of the problem is $Theta(log log n)$ rounds. On the other hand, there are also many LCLs in which randomness is useless. Previously, it was not known if there are any LCLs that benefit from randomness, but only subexponentially. We show that such problems exist: for example, there is an LCL with deterministic complexity $Theta(log^2 n)$ rounds and randomized complexity $Theta(log n log log n)$ rounds.
We analyze complex networks under random matrix theory framework. Particularly, we show that $Delta_3$ statistic, which gives information about the long range correlations among eigenvalues, provides a qualitative measure of randomness in networks. As networks deviate from the regular structure, $Delta_3$ follows random matrix prediction of linear behavior, in semi-logarithmic scale with the slope of $1/pi^2$, for the longer scale.