Large-Sized Network: Independent AV agents¶
Network Overview¶
In these notebooks, we utilize the Ingolstadt network within our simulator, SUMO. Since agents exhibit selfish behavior, we employ independent learning algorithms to model their decision-making.
Users can customize parameters for the
TrafficEnvironmentclass by consulting therouterl/environment/params.jsonfile. Based on its contents, they can create a dictionary with their preferred settings and pass it as an argument to theTrafficEnvironmentclass.
Included Tutorials:¶
IQL Tutorial. Uses Independent Q-Learning (IQL) (IQL) as an initial baseline for training decentralized policies.
IPPO Tutorial. Implements Independent Proximal Policy Optimization (IPPO) (IPPO), which has demonstrated strong benchmark performance in various tasks (paper1, paper2).
ISAC Tutorial. Uses Independent SAC (ISAC), the multi-agent extension of Soft Actor-Critic (SAC) (SAC), which balances exploration and exploitation using entropy-regularized reinforcement learning.
Ingolstadt Network Visualization¶