No Arabic abstract
This paper concerns applications of a recently-developed output-tracking technique to trajectory control of autonomous vehicles. The technique is based on three principles: Newton-Raphson flow for solving algebraic equations,output prediction, and controller speedup. Early applications of the technique, made to simple systems of an academic nature,were implemented by simple algorithms requiring modest computational efforts. In contrast, this paper tests it on commonly-used dynamic models to see if it can handle more complex control scenarios. Results are derived from simulations as well as a laboratory setting, and they indicate effective tracking convergence despite the simplicity of the control algorithm.
This paper presents a control technique for output tracking of reference signals in continuous-time dynamical systems. The technique is comprised of the following three elements: (i) output prediction which has to track the reference signal, (ii) a controller based on an integrator with variable gain, and (iii) a speedup of the control action for enhancing the trackers accuracy and, in some cases, guaranteeing stability of the closed-loop system. The technique is suitable for linear and nonlinear systems, implementable by simple algorithms, can track reference points as well as time-dependent reference signals, and may have large, even global domains of attraction. The derived theoretical results include convergence of the tracking controller and error analysis, and are supported by illustrative simulation and laboratory experiments.
Emergent cooperative adaptive cruise control (CACC) strategies being proposed in the literature for platoon formation in the Connected Autonomous Vehicle (CAV) context mostly assume idealized fixed information flow topologies (IFTs) for the platoon, implying guaranteed vehicle-to-vehicle (V2V) communications for the IFT assumed. Since CACC strategies entail continuous information broadcasting, communication failures can occur in congested CAV traffic networks, leading to a platoons IFT varying dynamically. To enhance the performance of CACC strategies, this study proposes the idea of dynamically optimizing the IFT for CACC, labeled the CACC-OIFT strategy. Under CACC-OIFT, the vehicles in the platoon cooperatively determine in real-time which vehicles will dynamically deactivate or activate the send functionality of their V2V communication devices to generate IFTs that optimize the platoon performance in terms of string stability under the ambient traffic conditions. Given the adaptive Proportional-Derivative (PD) controller with a two-predecessor-following scheme, and the ambient traffic conditions and the platoon size just before the start of a time period, the IFT optimization model determines the optimal IFT that maximizes the expected string stability. The optimal IFT is deployed for that time period, and the adaptive PD controller continuously determines the car-following behaviors of the vehicles based on the unfolding degeneration scenario for each time instant within that period. The effectiveness of the proposed CACC-OIFT is validated through numerical experiments in NS-3 based on NGSIM field data. The results indicate that the proposed CACC-OIFT can significantly enhance the string stability of platoon control in an unreliable V2V communication context, outperforming CACCs with fixed IFTs or with passive adaptive schemes for IFT dynamics.
Vehicle-to-vehicle communications can be unreliable as interference causes communication failures. Thereby, the information flow topology for a platoon of Connected Autonomous Vehicles (CAVs) can vary dynamically. This limits existing Cooperative Adaptive Cruise Control (CACC) strategies as most of them assume a fixed information flow topology (IFT). To address this problem, we introduce a CACC design that considers a dynamic information flow topology (CACC-DIFT) for CAV platoons. An adaptive Proportional-Derivative (PD) controller under a two-predecessor-following IFT is proposed to reduce the negative effects when communication failures occur. The PD controller parameters are determined to ensure the string stability of the platoon. Further, the designed controller also factors the performance of individual vehicles. Hence, when communication failure occurs, the system will switch to a certain type of CACC instead of degenerating to adaptive cruise control, which improves the control performance considerably. The effectiveness of the proposed CACC-DIFT is validated through numerical experiments based on NGSIM field data. Results indicate that the proposed CACC-DIFT design outperforms a CACC with a predetermined information flow topology.
Using deep reinforcement learning, we train control policies for autonomous vehicles leading a platoon of vehicles onto a roundabout. Using Flow, a library for deep reinforcement learning in micro-simulators, we train two policies, one policy with noise injected into the state and action space and one without any injected noise. In simulation, the autonomous vehicle learns an emergent metering behavior for both policies in which it slows to allow for smoother merging. We then directly transfer this policy without any tuning to the University of Delaware Scaled Smart City (UDSSC), a 1:25 scale testbed for connected and automated vehicles. We characterize the performance of both policies on the scaled city. We show that the noise-free policy winds up crashing and only occasionally metering. However, the noise-injected policy consistently performs the metering behavior and remains collision-free, suggesting that the noise helps with the zero-shot policy transfer. Additionally, the transferred, noise-injected policy leads to a 5% reduction of average travel time and a reduction of 22% in maximum travel time in the UDSSC. Videos of the controllers can be found at https://sites.google.com/view/iccps-policy-transfer.
This paper presents a novel model-reference reinforcement learning algorithm for the intelligent tracking control of uncertain autonomous surface vehicles with collision avoidance. The proposed control algorithm combines a conventional control method with reinforcement learning to enhance control accuracy and intelligence. In the proposed control design, a nominal system is considered for the design of a baseline tracking controller using a conventional control approach. The nominal system also defines the desired behaviour of uncertain autonomous surface vehicles in an obstacle-free environment. Thanks to reinforcement learning, the overall tracking controller is capable of compensating for model uncertainties and achieving collision avoidance at the same time in environments with obstacles. In comparison to traditional deep reinforcement learning methods, our proposed learning-based control can provide stability guarantees and better sample efficiency. We demonstrate the performance of the new algorithm using an example of autonomous surface vehicles.