Skip to main content

Coupling User Preference with External Rewards to Enable Driver-centered and Resource-aware EV Charging Recommendation

  • Conference paper
  • First Online:
Machine Learning and Knowledge Discovery in Databases (ECML PKDD 2022)

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 13716))

Abstract

Electric Vehicle (EV) charging recommendation that both accommodates user preference and adapts to the ever-changing external environment arises as a cost-effective strategy to alleviate the range anxiety of private EV drivers. Previous studies focus on centralized strategies to achieve optimized resource allocation, particularly useful for privacy-indifferent taxi fleets and fixed-route public transits. However, private EV driver seeks a more personalized and resource-aware charging recommendation that is tailor-made to accommodate the user preference (when and where to charge) yet sufficiently adaptive to the spatiotemporal mismatch between charging supply and demand. Here we propose a novel Regularized Actor-Critic (RAC) charging recommendation approach that would allow each EV driver to strike an optimal balance between the user preference (historical charging pattern) and the external reward (driving distance and wait time). Experimental results on two real-world datasets demonstrate the unique features and superior performance of our approach to the competing methods.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 79.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 99.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Notes

  1. 1.

    https://data.dundeecity.gov.uk/dataset/.

  2. 2.

    http://ubdc.gla.ac.uk/dataset/.

  3. 3.

    https://developers.google.com/maps/documentation/places/web-service.

References

  1. O’Donovan, C., Frith, J.: Electric buses in cities: driving towards cleaner air and lower co 2. Tech. Rep, Bloomberg New Energy Finance, NY, USA (2018)

    Google Scholar 

  2. Wang, X., Yuen, C., Hassan, N.U., An, N., Wu, W.: Electric vehicle charging station placement for urban public bus systems. IEEE Trans. Intell. Transp. Syst. 18(1), 128–139 (2016)

    Article  Google Scholar 

  3. Wang, G., Xie, X., Zhang, F., Liu, Y., Zhang, D.: Bcharge: data-driven real-time charging scheduling for large-scale electric bus fleets. In: RTSS, pp. 45–55 (2018)

    Google Scholar 

  4. Guo, T., You, P., Yang, Z.: Recommendation of geographic distributed charging stations for electric vehicles: a game theoretical approach. In: 2017 IEEE Power & Energy Society General Meeting, pp. 1–5 (2017)

    Google Scholar 

  5. Cao, Y., Kaiwartya, O., Zhuang, Y., Ahmad, N., Sun, Y., Lloret, J.: A decentralized deadline-driven electric vehicle charging recommendation. IEEE Syst. J. (2018)

    Google Scholar 

  6. Qiang, Y., Li, C., Brocanelli, M., Zhu, D.: Counterfactual interpolation augmentation (CIA): a unified approach to enhance fairness and explainability of DNN. In: IJCAI (2022)

    Google Scholar 

  7. Li, X., Li, X., Pan, D., Zhu, D.: Improving adversarial robustness via probabilistically compact loss with logit constraints. In: AAAI, vol. 35, pp. 8482–8490 (2021)

    Google Scholar 

  8. Pan, D., Li, X., Zhu, D.: Explaining deep neural network models with adversarial gradient integration. In: IJCAI, pp. 2876–2883 (2021)

    Google Scholar 

  9. Pan, D., Li, X., Li, X., Zhu, D.: Explainable recommendation via interpretable feature mapping and evaluation of explainability. In: IJCAI, pp. 2690–2696 (2020)

    Google Scholar 

  10. Zhao, K., et al.: Discovering subsequence patterns for next poi recommendation. In: IJCAI, pp. 3216–3222 (2020)

    Google Scholar 

  11. Rendle, S., Freudenthaler, C., Schmidt-Thieme, L.: Factorizing personalized markov chains for next-basket recommendation. In: WWW 2010, pp. 811–820 (2010)

    Google Scholar 

  12. Zhu, Y., et al.: What to do next: modeling user behaviors by time-LSTM. In: IJCAI, vol. 17, pp. 3602–3608 (2017)

    Google Scholar 

  13. Kong, D., Wu, F.: HST-LSTM: a hierarchical spatial-temporal long-short term memory network for location prediction. In: IJCAI, vol. 18, pp. 2341–2347 (2018)

    Google Scholar 

  14. Zhao, P., et al.: Where to go next: a spatio-temporal gated network for next poi recommendation. In: AAAI, vol. 33, pp. 5877–5884 (2019)

    Google Scholar 

  15. Li, L., Zheng, L., Yang, F., Li, T.: Modeling and broadening temporal user interest in personalized news recommendation. Expert Syst. Appl. 41(7), 3168–3177 (2014)

    Article  Google Scholar 

  16. Liu, Q., Wu, S., Wang, L., Tan, T.: Predicting the next location: a recurrent model with spatial and temporal contexts. In: AAAI, vol. 30 (2016)

    Google Scholar 

  17. Huang, L., Ma, Y., Wang, S., Liu, Y.: An attention-based spatiotemporal LSTM network for next poi recommendation. IEEE Trans. Serv. Comput. (2019)

    Google Scholar 

  18. Wu, Y., Li, K., Zhao, G., Qian, X.: Long-and short-term preference learning for next poi recommendation. In: CIKM 2019, pp. 2301–2304 (2019)

    Google Scholar 

  19. Wang, E., et al.: Joint charging and relocation recommendation for e-taxi drivers via multi-agent mean field hierarchical reinforcement learning. IEEE Trans. Mobile Comput. (2020)

    Google Scholar 

  20. Zhang, W., et al.: Intelligent electric vehicle charging recommendation based on multi-agent reinforcement learning. In: WWW 2021, pp. 1856–1867 (2021)

    Google Scholar 

  21. Massimo, D., Ricci, F.: Harnessing a generalised user behaviour model for next-poi recommendation. In: RecSys 2018, pp. 402–406 (2018)

    Google Scholar 

  22. Zheng, G., et al.: DRN: a deep reinforcement learning framework for news recommendation. In: WWW 2018 (2018)

    Google Scholar 

  23. Wang, L., Zhang, W., He, X., Zha, H.: Supervised reinforcement learning with recurrent neural network for dynamic treatment recommendation. In: KDD 2018, pp. 2447–2456 (2018)

    Google Scholar 

  24. Watkins, C.J., Dayan, P.: Q-learning. Mach. Learn. 8(3–4), 279–292 (1992)

    Google Scholar 

  25. Sutton, R.S., McAllester, D.A., Singh, S.P., Mansour, Y., et al.: Policy gradient methods for reinforcement learning with function approximation. In: NIPS, vol. 99, pp. 1057–1063 (1999)

    Google Scholar 

  26. Schmitt, S., Hessel, M., Simonyan, K.: Off-policy actor-critic with shared experience replay. In: ICML (2020)

    Google Scholar 

Download references

Acknowledgements

This work is supported by the National Science Foundation under grant no. IIS-1724227.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Dongxiao Zhu .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Li, C., Dong, Z., Fisher, N., Zhu, D. (2023). Coupling User Preference with External Rewards to Enable Driver-centered and Resource-aware EV Charging Recommendation. In: Amini, MR., Canu, S., Fischer, A., Guns, T., Kralj Novak, P., Tsoumakas, G. (eds) Machine Learning and Knowledge Discovery in Databases. ECML PKDD 2022. Lecture Notes in Computer Science(), vol 13716. Springer, Cham. https://doi.org/10.1007/978-3-031-26412-2_1

Download citation

  • DOI: https://doi.org/10.1007/978-3-031-26412-2_1

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-031-26411-5

  • Online ISBN: 978-3-031-26412-2

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics