Learning Coordinated Performant Flight with 20 Neurons
Author(s)
Sukhatme, Gaurav
Advisor(s)
Editor(s)
Collections
Supplementary to:
Permanent Link
Abstract
We have recently demonstrated the possibility of learning controllers that are zero-shot transferable to groups of real quadrotors via large-scale, multi-agent, end-to-end reinforcement learning. We train policies parameterized by neural networks that can control individual drones in a group in a fully decentralized manner. Our policies, trained in simulated environments with realistic quadrotor physics, demonstrate advanced flocking behaviors, perform aggressive maneuvers in tight formations while avoiding collisions with each other, break and re-establish formations to avoid collisions with moving obstacles, and efficiently coordinate in pursuit-evasion tasks. The model learned in simulation transfers to highly resource-constrained physical quadrotors. Motivated by these results and the observation that neural control of memory-constrained, agile robots requires small yet highly performant models, the talk will conclude with some thoughts on coaxing learned models onto devices with modest computational capabilities.
Sponsor
Date
2025-04-09
Extent
59:39 minutes
Resource Type
Moving Image
Resource Subtype
Lecture
Rights Statement
Unless otherwise noted, all materials are protected under U.S. Copyright Law and all rights are reserved