Large-Sized Network: Independent AV agents¶
Network Overview¶
In these notebooks, we utilize the Ingolstadt network within our simulator, SUMO. Since agents exhibit selfish behavior, we employ independent learning algorithms to model their decision-making.
Users can customize parameters for the
TrafficEnvironment
class by consulting therouterl/environment/params.json
file. Based on its contents, they can create a dictionary with their preferred settings and pass it as an argument to theTrafficEnvironment
class.
Included Tutorials:¶
IQL Tutorial. Uses Independent Q-Learning (IQL) (IQL) as an initial baseline for training decentralized policies.
IPPO Tutorial. Implements Independent Proximal Policy Optimization (IPPO) (IPPO), which has demonstrated strong benchmark performance in various tasks (paper1, paper2).
ISAC Tutorial. Uses Independent SAC (ISAC), the multi-agent extension of Soft Actor-Critic (SAC) (SAC), which balances exploration and exploitation using entropy-regularized reinforcement learning.
Ingolstadt Network Visualization¶