Paper ID: 2504.06125 • Published Apr 8, 2025
Robo-taxi Fleet Coordination at Scale via Reinforcement Learning
Luigi Tresca, Carolin Schmidt, James Harrison, Filipe Rodrigues, Gioele Zardini, Daniele Gammelli, Marco Pavone
Politecnico di Torino•Technical University of Denmark•Google DeepMind•MIT•Stanford University
TL;DR
Get AI-generated summaries with premium
Get AI-generated summaries with premium
Fleets of robo-taxis offering on-demand transportation services, commonly
known as Autonomous Mobility-on-Demand (AMoD) systems, hold significant promise
for societal benefits, such as reducing pollution, energy consumption, and
urban congestion. However, orchestrating these systems at scale remains a
critical challenge, with existing coordination algorithms often failing to
exploit the systems' full potential. This work introduces a novel
decision-making framework that unites mathematical modeling with data-driven
techniques. In particular, we present the AMoD coordination problem through the
lens of reinforcement learning and propose a graph network-based framework that
exploits the main strengths of graph representation learning, reinforcement
learning, and classical operations research tools. Extensive evaluations across
diverse simulation fidelities and scenarios demonstrate the flexibility of our
approach, achieving superior system performance, computational efficiency, and
generalizability compared to prior methods. Finally, motivated by the need to
democratize research efforts in this area, we release publicly available
benchmarks, datasets, and simulators for network-level coordination alongside
an open-source codebase designed to provide accessible simulation platforms and
establish a standardized validation process for comparing methodologies. Code
available at: this https URL
Figures & Tables
Unlock access to paper figures and tables to enhance your research experience.