Abstract
Traditional public transportation mostly uses fixed schedules to meet society’s needs reliably. Alternatively, one can design a dynamic policy to update the schedule according to the state of the network. Using such policy, any dynamic changes in demand, travel times, traffic, geometry, or emergency modes will be reflected in the real-time scheduling. In this study, the application of reinforcement learning for public transportation scheduling and routing is proposed considering electric vehicles. Reinforcement learning is a right choice for online decision-making in a time-variant setting. Reflecting electric vehicles’ characteristics on public transportation scheduling supports the necessity of considering a dynamic scheduling policy, such as vehicle-to-grid transaction capability or dynamic charging strategies. The proposed scheduling methodology is evaluated for a customary six-stop network. The trained agent takes actions based on the number of waiting people, the drop-off requests, battery level, time, and the current location. The assumed models for the traffic load effect, the passenger flow per stops, and the electricity price add nonlinearity to the problem in which this algorithm will behave more conveniently compared to a fixed schedule setting. Using this method, the electric bus can choose the next destination, the optimal driving speed for each interval, and the amount of electricity to charge or to sell for optimal financial gain when connected to the charging infrastructure. The proposed environment is built numerically, and a learning-based agent is trained for this purpose. Our study supports that the optimal dynamic agent outperforms a fixed schedule policy in different modes of network performance, particularly, when the network undergoes abrupt demand changes.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Laporte, G.: The vehicle routing problem: an overview of exact and approximate algorithms. Eur. J. Oper. Res. 59(3), 345–358 (1992)
Pillac, V., Gendreau, M., Guéret, C., Medaglia, A.L.: A review of dynamic vehicle routing problems. Eur. J. Oper. Res. 225(1), 1–11 (2013)
Dijkstra, E.W., et al.: A note on two problems in connexion with graphs. Numer. Math. 1(1), 269–271 (1959)
Hart, P.E., Nilsson, N.J., Raphael, B.: A formal basis for the heuristic determination of minimum cost paths. IEEE Trans. Syst. Sci. Cybernetics. 4(2), 100–107 (1968)
Nilsson, N.J.: Principles of Artificial Intelligence. Morgan Kaufmann, San Francisco (2014)
Gelperin, D.: On the optimality of A*. Artif. Intell. 8(1), 69–76 (1977)
Bellman, R.E.: Dynamic Programming, p. 151. Princeton University Press. Bellman Dynamic Programming., Princeton (1957)
Ferguson, D., Likhachev, M., Stentz, A.: A guide to heuristic-based path planning. In Proceedings of the international workshop on planning under uncertainty for autonomous systems, International Conference on Automated Planning and Scheduling (ICAPS), pp. 9–18 (2005)
Russell, S. and P. Norvig, Artificial intelligence: a modern approach, 2002
Regan, A.C., Mahmassani, H.S., Jaillet, P.: Dynamic decision making for commercial fleet operations using real-time information. Transp. Res. Rec. 1537(1), 91–97 (1996)
Yang, J., Jaillet, P., Mahmassani, H.S.: On-line algorithms for truck fleet assignment and scheduling under real-time information. Transp. Res. Rec. 1667(1), 107–113 (1999)
Liao, T.-Y.: On-line vehicle routing problems for carbon emissions reduction. Comput. Aided Civ. Inf. Eng. 32(12), 1047–1063 (2017)
Hu, T.-Y.: Evaluation framework for dynamic vehicle routing strategies under real-time information. Transp. Res. Rec. 1774(1), 115–122 (2001)
Ichoua, S., Gendreau, M., Potvin, J.-Y.: Vehicle dispatching with time-dependent travel times. Eur. J. Oper. Res. 144(2), 379–396 (2003)
Liao, T.-Y., Wang, S.-J., Hu, T.-Y.: On-line vehicle routing problems: a hybrid metaheuristic approach. J. East. Asia Soc. Transp. Stud. 9, 660–675 (2011)
Barkaoui, M., Gendreau, M.: An adaptive evolutionary approach for real-time vehicle routing and dispatching. Comput. Oper. Res. 40(7), 1766–1776 (2013)
Schneider, M., Stenger, A., Goeke, D.: The electric vehicle-routing problem with time windows and recharging stations. Transp. Sci. 48(4), 500–520 (2014)
Sutton, R.S., Barto, A.G.: Reinforcement Learning: an Introduction. MIT press, Cambridge, MA (2018)
Guille, C., Gross, G.: A conceptual framework for the vehicle-to-grid (V2G) implementation. Energy Policy. 37(11), 4379–4390 (2009)
Khoshmagham, S., P. Hosseini, S. Perley, T. Barkley, A. Prasad.: Travel Time and Flow Prediction Using a Machine Learning Algorithm to Optimize Traffic Responsive Coordination Plans: Santa Clara County Expressways. In Transportation Research Board Annual Meeting (2019)
Haarnoja, T., Zhou, A., Abbeel, P., Levine, S.: Soft actor-critic: Off-policy maximum entropy deep reinforcement learning with a stochastic actor. arXiv preprint arXiv:1801.01290 (2018)
Haarnoja, T., Zhou, A., Hartikainen, K., Tucker, G., Ha, S., Tan, J., Kumar, V., Zhu, H., Gupta, A., Abbeel, P. et al. Soft actor-critic algorithms and applications. arXiv preprint arXiv:1812.05905 (2018)
Haarnoja, T., Ha, S., Zhou, A., Tan, J., Tucker, G., Levine, S.: Learning to walk via deep reinforcement learning. arXiv preprint arXiv:1812.11103 (2018)
Acknowledgments
Research funding is partially provided by a grant from the US Department of Transportation’s University Transportation Centers Program and National Science Foundation through Grants CMMI-1351537 and by grants from the Commonwealth of Pennsylvania; Department of Community and Economic Development, through the Pennsylvania Infrastructure Technology Alliance (PITA); and Center for Integrated Asset Management for MultiModal Transportation Infrastructure Systems (CIAMTIS).
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2022 The Society for Experimental Mechanics, Inc.
About this paper
Cite this paper
Eshkevari, S.S., Eshkevari, S.S., Pakzad, S.N., Muñoz-Avila, H., Kishore, S. (2022). Routing of Public and Electric Transportation Systems Using Reinforcement Learning. In: Madarshahian, R., Hemez, F. (eds) Data Science in Engineering, Volume 9. Conference Proceedings of the Society for Experimental Mechanics Series. Springer, Cham. https://doi.org/10.1007/978-3-030-76004-5_31
Download citation
DOI: https://doi.org/10.1007/978-3-030-76004-5_31
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-76003-8
Online ISBN: 978-3-030-76004-5
eBook Packages: EngineeringEngineering (R0)