Do you want to publish a course? Click here

Field Estimation using Robotic Swarms through Bayesian Regression and Mean-Field Feedback

251   0   0.0 ( 0 )
 Added by Tongjia Zheng
 Publication date 2021
and research's language is English




Ask ChatGPT about the research

Recent years have seen an increased interest in using mean-field density based modelling and control strategy for deploying robotic swarms. In this paper, we study how to dynamically deploy the robots subject to their physical constraints to efficiently measure and reconstruct certain unknown spatial field (e.g. the air pollution index over a city). Specifically, the evolution of the robots density is modelled by mean-field partial differential equations (PDEs) which are uniquely determined by the robots individual dynamics. Bayesian regression models are used to obtain predictions and return a variance function that represents the confidence of the prediction. We formulate a PDE constrained optimization problem based on this variance function to dynamically generate a reference density signal which guides the robots to uncertain areas to collect new data, and design mean-field feedback-based control laws such that the robots density converges to this reference signal. We also show that the proposed feedback law is robust to density estimation errors in the sense of input-to-state stability. Simulations are included to verify the effectiveness of the algorithms.



rate research

Read More

237 - Tongjia Zheng , Qing Han , 2020
With the rapid development of AI and robotics, transporting a large swarm of networked robots has foreseeable applications in the near future. Existing research in swarm robotics has mainly followed a bottom-up philosophy with predefined local coordination and control rules. However, it is arduous to verify the global requirements and analyze their performance. This motivates us to pursue a top-down approach, and develop a provable control strategy for deploying a robotic swarm to achieve a desired global configuration. Specifically, we use mean-field partial differential equations (PDEs) to model the swarm and control its mean-field density (i.e., probability density) over a bounded spatial domain using mean-field feedback. The presented control law uses density estimates as feedback signals and generates corresponding velocity fields that, by acting locally on individual robots, guide their global distribution to a target profile. The design of the velocity field is therefore centralized, but the implementation of the controller can be fully distributed -- individual robots sense the velocity field and derive their own velocity control signals accordingly. The key contribution lies in applying the concept of input-to-state stability (ISS) to show that the perturbed closed-loop system (a nonlinear and time-varying PDE) is locally ISS with respect to density estimation errors. The effectiveness of the proposed control laws is verified using agent-based simulations.
131 - Tongjia Zheng , Qing Han , Hai Lin 2021
Swarm robotic systems have foreseeable applications in the near future. Recently, there has been an increasing amount of literature that employs mean-field partial differential equations (PDEs) to model the time-evolution of the probability density of swarm robotic systems and uses mean-field feedback to design stable control laws that act on individuals such that their density converges to a target profile. However, it remains largely unexplored considering problems of how to estimate the mean-field density, how the density estimation algorithms affect the control performance, and whether the estimation performance in turn depends on the control algorithms. In this work, we focus on studying the interplay of these algorithms. Specially, we propose new mean-field control laws which use the real-time density and its gradient as feedback, and prove that they are globally input-to-state stable (ISS) to estimation errors. Then, we design filtering algorithms to obtain estimates of the density and its gradient, and prove that these estimates are convergent assuming the control laws are known. Finally, we show that the feedback interconnection of these estimation and control algorithms is still globally ISS, which is attributed to the bilinearity of the mean-field PDE system. An agent-based simulation is included to verify the stability of these algorithms and their feedback interconnection.
This work studies how to estimate the mean-field density of large-scale systems in a distributed manner. Such problems are motivated by the recent swarm control technique that uses mean-field approximations to represent the collective effect of the swarm, wherein the mean-field density (and its gradient) is usually used in feedback control design. In the first part, we formulate the density estimation problem as a filtering problem of the associated mean-field partial differential equation (PDE), for which we employ kernel density estimation (KDE) to construct noisy observations and use filtering theory of PDE systems to design an optimal (centralized) density filter. It turns out that the covariance operator of observation noise depends on the unknown density. Hence, we use approximations for the covariance operator to obtain a suboptimal density filter, and prove that both the density estimates and their gradient are convergent and remain close to the optimal one using the notion of input-to-state stability (ISS). In the second part, we continue to study how to decentralize the density filter such that each agent can estimate the mean-field density based on only its own position and local information exchange with neighbors. We prove that the local density filter is also convergent and remains close to the centralized one in the sense of ISS. Simulation results suggest that the (centralized) suboptimal density filter is able to generate convergent density estimates, and the local density filter is able to converge and remain close to the centralized filter.
127 - Xuehe Wang , Lingjie Duan 2020
Today many mobile users in various zones are invited to sense and send back real-time useful information (e.g., traffic observation and sensor data) to keep the freshness of the content updates in such zones. However, due to the sampling cost in sensing and transmission, a user may not have the incentive to contribute the real-time information to help reduce the age of information (AoI). We propose dynamic pricing for each zone to offer age-dependent monetary returns and encourage users to sample information at different rates over time. This dynamic pricing design problem needs to well balance the monetary payments as rewards to users and the AoI evolution over time, and is challenging to solve especially under the incomplete information about users arrivals and their private sampling costs. After formulating the problem as a nonlinear constrained dynamic program, to avoid the curse of dimensionality, we first propose to approximate the dynamic AoI reduction as a time-average term and successfully solve the approximate dynamic pricing in closed-form. Further, by providing the steady-state analysis for an infinite time horizon, we show that the pricing scheme (though in closed-form) can be further simplified to an $varepsilon$-optimal version without recursive computing over time. Finally, we extend the AoI control from a single zone to many zones with heterogeneous user arrival rates and initial ages, where each zone cares not only its own AoI dynamics but also the average AoI of all the zones in a mean field game system to provide a holistic service. Accordingly, we propose decentralized mean field pricing for each zone to self-operate by using a mean field term to estimate the average age dynamics of all the zones, which does not even require many zones to exchange their local data with each other.
In this paper, we consider discrete-time partially observed mean-field games with the risk-sensitive optimality criterion. We introduce risk-sensitivity behaviour for each agent via an exponential utility function. In the game model, each agent is weakly coupled with the rest of the population through its individual cost and state dynamics via the empirical distribution of states. We establish the mean-field equilibrium in the infinite-population limit using the technique of converting the underlying original partially observed stochastic control problem to a fully observed one on the belief space and the dynamic programming principle. Then, we show that the mean-field equilibrium policy, when adopted by each agent, forms an approximate Nash equilibrium for games with sufficiently many agents. We first consider finite-horizon cost function, and then, discuss extension of the result to infinite-horizon cost in the next-to-last section of the paper.
comments
Fetching comments Fetching comments
Sign in to be able to follow your search criteria
mircosoft-partner

هل ترغب بارسال اشعارات عن اخر التحديثات في شمرا-اكاديميا