TY - JOUR

T1 - Controllability and Stabilization for Herding a Robotic Swarm Using a Leader

T2 - A Mean-Field Approach

AU - Elamvazhuthi, Karthik

AU - Kakish, Zahi

AU - Shirsat, Aniket

AU - Berman, Spring

N1 - Funding Information:
Manuscript received April 25, 2020; accepted September 15, 2020. Date of publication October 28, 2020; date of current version April 2, 2021. This work was supported by ONR Young Investigator Award N00014-16-1-2605, and in part by the Arizona State University Global Security Initiative. This article was recommended for publication by Associate Editor M. Schwager and Editor P. Robuffo upon evaluation of the reviewers’ comments. (Corresponding author: Karthik Elamvazhuthi.) Karthik Elamvazhuthi is with the Department of Mathematics, University of California, Los Angeles, CA 90095 USA (e-mail: karthikevaz@math.ucla.edu).
Publisher Copyright:
© 2004-2012 IEEE.

PY - 2021/4

Y1 - 2021/4

N2 - In this article, we introduce a model and a control approach for herding a swarm of 'follower' agents to a target distribution among a set of states using a single 'leader' agent. The follower agents evolve on a finite state space that is represented by a graph and transition between states according to a continuous-time Markov chain (CTMC), whose transition rates are determined by the location of the leader agent. The control problem is to define a sequence of states for the leader agent that steers the probability density of the forward equation of the Markov chain. For the case, when the followers are possibly interacting, we prove local approximate controllability of the system about equilibrium probability distributions. For the case, when the followers are noninteracting, we design two switching control laws for the leader that drive the swarm of follower agents asymptotically to a target probability distribution that is positive for all states. The first strategy is open-loop in nature, and the switching times of the leader are independent of the follower distribution. The second strategy is of feedback type, and the switching times of the leader are functions of the follower density in the leader's current state. We validate our control approach through numerical simulations with varied numbers of follower agents that evolve on graphs of different sizes, through a 3-D multirobot simulation in which a quadrotor is used to control the spatial distribution of eight ground robots over four regions, and through a physical experiment in which a swarm of ten robots is herded by a virtual leader over four regions.

AB - In this article, we introduce a model and a control approach for herding a swarm of 'follower' agents to a target distribution among a set of states using a single 'leader' agent. The follower agents evolve on a finite state space that is represented by a graph and transition between states according to a continuous-time Markov chain (CTMC), whose transition rates are determined by the location of the leader agent. The control problem is to define a sequence of states for the leader agent that steers the probability density of the forward equation of the Markov chain. For the case, when the followers are possibly interacting, we prove local approximate controllability of the system about equilibrium probability distributions. For the case, when the followers are noninteracting, we design two switching control laws for the leader that drive the swarm of follower agents asymptotically to a target probability distribution that is positive for all states. The first strategy is open-loop in nature, and the switching times of the leader are independent of the follower distribution. The second strategy is of feedback type, and the switching times of the leader are functions of the follower density in the leader's current state. We validate our control approach through numerical simulations with varied numbers of follower agents that evolve on graphs of different sizes, through a 3-D multirobot simulation in which a quadrotor is used to control the spatial distribution of eight ground robots over four regions, and through a physical experiment in which a swarm of ten robots is herded by a virtual leader over four regions.

KW - Feedback control

KW - herding

KW - Markov chains

KW - multirobot systems

KW - stabilization

KW - swarms

KW - switching systems

UR - http://www.scopus.com/inward/record.url?scp=85103930991&partnerID=8YFLogxK

UR - http://www.scopus.com/inward/citedby.url?scp=85103930991&partnerID=8YFLogxK

U2 - 10.1109/TRO.2020.3031237

DO - 10.1109/TRO.2020.3031237

M3 - Article

AN - SCOPUS:85103930991

VL - 37

SP - 418

EP - 432

JO - IEEE Transactions on Robotics

JF - IEEE Transactions on Robotics

SN - 1552-3098

IS - 2

M1 - 9241729

ER -