skip to main content


This content will become publicly available on May 29, 2024

Title: Multi-Agent Spatial Predictive Control with Application to Drone Flocking
We introduce Spatial Predictive Control (SPC), a technique for solving the following problem: given a collection of robotic agents with black-box positional low-level controllers (PLLCs) and a mission-specific distributed cost function, how can a distributed controller achieve and maintain cost-function minimization without a plant model and only positional observations of the environment? Our fully distributed SPC controller is based strictly on the position of the agent itself and on those of its neighboring agents. This information is used in every time step to compute the gradient of the cost function and to perform a spatial look-ahead to predict the best next target position for the PLLC. Using a simulation environment, we show that SPC outperforms Potential Field Controllers, a related class of controllers, on the drone flocking problem. We also show that SPC works on real hardware, and is therefore able to cope with the potential sim-to-real transfer gap. We demonstrate its performance using as many as 16 Crazyflie 2.1 drones in a number of scenarios, including obstacle avoidance.  more » « less
Award ID(s):
1918225
NSF-PAR ID:
10432250
Author(s) / Creator(s):
; ; ; ;
Date Published:
Journal Name:
Proceedings of ICRA 2023. IEEE International Conference on Robotics and Automation
Page Range / eLocation ID:
1221 to 1227
Format(s):
Medium: X
Sponsoring Org:
National Science Foundation
More Like this
  1. Hardware-in-the-loop (HIL) testing is a popular control system testing method because it bridges the gap between modeling/simulation and experiments. Instead of designing a full hardware-based experiment to validate the results of a simulation, the plant hardware can be replaced with an emulator device that responds to exogenous inputs and effectively emulates the dynamic behavior of a system. This approach can be more cost-effective and modular, since the emulated plant system can be modeled in a simulation environment, implemented on a simplified piece of hardware and changed quickly without having to fabricate new parts. This paper develops the hardware and control scheme for a certain type of HIL device called a dynamic load emulator that consists of a 1-DOF linear hydraulic dynamometer equipped with in-line sensing to measure both its own position and the force exerted on it by a device-under-test. This measured force is passed to a real-time model of the emulated dynamic system. The model outputs the emulated system position, and a closed-loop controller is used to emulate this position. The emulator controller incorporates both model-based feedforward and standard feedback PI control. This paper characterizes the dynamometer-based dynamic load emulator and its controller, determining its hardware limitations and validating its capabilities when experiencing a force input from a linear spring with known parameters. Additionally, this paper demonstrates the ability of the emulator to represent the dynamics of a 1-DOF robotic joint when actuated by a pair of fluidic artificial muscles (FAMs). The primary contribution of this work is to allow for more comprehensive testing of FAM configurations, topologies, and controllers for a wide range of parameters, because the same hardware can be used to emulate multiple systems. As a result, this work will lead to more cost-effective, time-efficient, and energy-efficient designs of robotic systems and the FAMs used to actuate them.

     
    more » « less
  2. We propose a new algorithm to simplify the controller development for distributed robotic systems subject to external observations, disturbances, and communication delays. Unlike prior approaches that propose specialized solutions to handling communication latency for specific robotic applications, our algorithm uses an arbitrary centralized controller as the specification and automatically generates distributed controllers with communication management and delay compensation. We formulate our goal as nonlinear optimal control— using a regret minimizing objective that measures how much the distributed agents behave differently from the delay-free centralized response—and solve for optimal actions w.r.t. local estimations of this objective using gradient-based optimization. We analyze our proposed algorithm’s behavior under a linear time-invariant special case and prove that the closed-loop dynamics satisfy a form of input-to-state stability w.r.t. unexpected disturbances and observations. Our experimental results on both simulated and real-world robotic tasks demonstrate the practical usefulness of our approach and show significant improvement over several baseline approaches. 
    more » « less
  3. We propose a new algorithm to simplify the controller development for distributed robotic systems subject to external observations, disturbances, and communication delays. Unlike prior approaches that propose specialized solutions to handling communication latency for specific robotic applications, our algorithm uses an arbitrary centralized controller as the specification and automatically generates distributed controllers with communication management and delay compensation. We formulate our goal as nonlinear optimal control— using a regret minimizing objective that measures how much the distributed agents behave differently from the delay-free centralized response—and solve for optimal actions w.r.t. local estimations of this objective using gradient-based optimization. We analyze our proposed algorithm’s behavior under a linear time-invariant special case and prove that the closed-loop dynamics satisfy a form of input-to-state stability w.r.t. unexpected disturbances and observations. Our experimental results on both simulated and real-world robotic tasks demonstrate the practical usefulness of our approach and show significant improvement over several baseline approaches. 
    more » « less
  4. We consider the problem of optimal control of district cooling energy plants (DCEPs) consisting of multiple chillers, a cooling tower, and a thermal energy storage (TES), in the presence of time-varying electricity price. A straightforward application of model predictive control (MPC) requires solving a challenging mixed-integer nonlinear program (MINLP) because of the on/off of chillers and the complexity of the DCEP model. Reinforcement learning (RL) is an attractive alternative since its real-time control computation is much simpler. But designing an RL controller is challenging due to myriad design choices and computationally intensive training. In this paper, we propose an RL controller and an MPC controller for minimizing the electricity cost of a DCEP and compare them via simulations. The two controllers are designed to be comparable in terms of objective and information requirements. The RL controller uses a novel Q-learning algorithm that is based on least-squares policy iteration. We describe the design choices for the RL controller, including the choice of state space and basis functions, that are found to be effective. The proposed MPC controller does not need a mixed integer solver for implementation, but only a nonlinear program (NLP) solver. A rule-based baseline controller is also proposed to aid in comparison. Simulation results show that the proposed RL and MPC controllers achieve similar savings over the baseline controller, about 17%. 
    more » « less
  5. Abstract This paper presents a hierarchical nonlinear control algorithm for the real-time planning and control of cooperative locomotion of legged robots that collaboratively carry objects. An innovative network of reduced-order models subject to holonomic constraints, referred to as interconnected linear inverted pendulum (LIP) dynamics, is presented to study cooperative locomotion. The higher level of the proposed algorithm employs a supervisory controller, based on event-based model predictive control (MPC), to effectively compute the optimal reduced-order trajectories for the interconnected LIP dynamics. The lower level of the proposed algorithm employs distributed nonlinear controllers to reduce the gap between reduced- and full-order complex models of cooperative locomotion. In particular, the distributed controllers are developed based on quadratic programing (QP) and virtual constraints to impose the full-order dynamical models of each agent to asymptotically track the reduced-order trajectories while having feasible contact forces at the leg ends. The paper numerically investigates the effectiveness of the proposed control algorithm via full-order simulations of a team of collaborative quadrupedal robots, each with a total of 22 degrees-of-freedom. The paper finally investigates the robustness of the proposed control algorithm against uncertainties in the payload mass and changes in the ground height profile. Numerical studies show that the cooperative agents can transport unknown payloads whose masses are up to 57%, 97%, and 137% of a single agent's mass with a team of two, three, and four legged robots. 
    more » « less