Tay, Bee Sim (2023) Public transport route optimization with reinforcement learning. Final Year Project, UTAR.
Abstract
High transportation demand due to a large population has resulted in traffic congestion problems in cities, which can be addressed through public transport. However, unbalanced passenger demands and traffic conditions can affect the performance of buses. The stop-skipping strategy effectively distributes passenger demand while minimizing bus operating costs if the operator can adapt to changes in passenger demands and traffic conditions. Therefore, this project proposes a deep reinforcement learning-based public transport route optimization where the agent can acquire the optimal strategy by interacting with the dynamic bus environment. This project aims to maximize the passenger satisfaction levels while minimizing bus operator expenditures. Thus, the dynamic bus environment is designed based on a bus optimization scheme that comprises one express bus followed by one no-skip bus to serve stranded passengers due to skipped actions. The reward function is formulated as a function of passenger demand, in-vehicle time, bus running time and passenger waiting time. which is used to train the double deep Q-network (DDQN) agent. Simulation results show that the agent can intelligently skip stations and outperform the conventional method under different passenger distribution patterns. The DDQN approach yields the best performance in the static passenger demand scenario, followed by the scenario with dynamic passenger demands according to time, and lastly the randomly distributed passenger demand scenario. Future studies should consider the load constraints of buses and other factors, such as bus utilization rate, to improve the performance of stop-skipping services for passengers and operators. Real-life passenger data could be incorporated into the DRL model using Internet of Things technology (IoT) for route optimization.
Actions (login required)