Abstract
Mobile crowdsourcing (MC) which has been developed rapidly in recent years is playing an increasingly indispensable role in people’s daily lives such as taxi-hailing, food delivery and other services. The geographic equilibrium of service supply and demand is crucial so that the MC system could guarantee more promising matches in a more regionally balanced way. However, due to the spatial dynamic of MC environments, the emergence of supply and demand is unpredictable, asymmetric, and constantly changing among different regions and throughout the day, presenting considerable challenges to the MC platform. In this paper, we propose a hybrid reinforcement learning and transformer-based balancing framework (HRB) to achieve geographically balanced coverage of MC services, considering both the imbalanced state of service supply-demand geographical distribution and the moving willingness of MC participants. The HRB framework is developed based on the Deep Deterministic Policy Gradient strategy, which includes an actor-critic network for generating migration strategies and a Willingness Transformer (WiT) model for predicting the migration willingness of both mobile service providers and demanders among different regions. Experimental results have validated the effectiveness by comparing the proposed approach with other algorithms under multiple indicators.
Keywords
- Mobile crowdsourcing service
- Supply and demand balance
- Reinforcement learning
- Transformer
- Migration willingness
This is a preview of subscription content, access via your institution.
Buying options





References
Chen, L., et al.: Dynamic cluster-based over-demand prediction in bike sharing systems. In: Proceedings of the 2016 ACM International Joint Conference on Pervasive and Ubiquitous Computing, UbiComp 2016, Heidelberg, Germany, 12–16 September2016. pp. 841–852. ACM (2016)
Covington, P., Adams, J., Sargin, E.: Deep neural networks for YouTube recommendations. In: Proceedings of the 10th ACM Conference on Recommender Systems, Boston, MA, USA, 15–19 September 2016. pp. 191–198. ACM (2016)
Dietterich, T.G.: Hierarchical reinforcement learning with the MAXQ value function decomposition. J. Artif. Intell. Res. 13, 227–303 (2000)
Dosovitskiy, A., et al.: An image is worth 16x16 words: transformers for image recognition at scale. In: 9th International Conference on Learning Representations, ICLR 2021, Virtual Event, Austria, 3–2 May 2021. OpenReview.net (2021)
Duan, Y., Wu, J.: Optimizing rebalance scheme for dock-less bike sharing systems with adaptive user incentive. In: 20th IEEE International Conference on Mobile Data Management, MDM 2019, Hong Kong, SAR, China, 10–13 June 2019. pp. 176–181. IEEE (2019)
Duan, Y., Wu, J.: Optimizing the crowdsourcing-based bike station rebalancing scheme. In: 39th IEEE International Conference on Distributed Computing Systems, ICDCS 2019, Dallas, TX, USA, 7–10 July 2019, pp. 1559–1568. IEEE (2019)
Hamrouni, A., Alelyani, T., Ghazzai, H., Massoud, Y.: Toward collaborative mobile crowdsourcing. IEEE Internet Things Mag. 4(2), 88–94 (2021)
He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2016, Las Vegas, NV, USA, 27–30 June 2016. pp. 770–778. IEEE Computer Society (2016)
Holler, J., et al.: Deep reinforcement learning for multi-driver vehicle dispatching and repositioning problem. In: 2019 IEEE International Conference on Data Mining, ICDM 2019, Beijing, China, 8–11 November 2019. pp. 1090–1095. IEEE (2019)
Jiao, Y., et al.: Real-world ride-hailing vehicle repositioning using deep reinforcement learning. CoRR abs/2103.04555 (2021)
Li, M., et al.: Efficient ridesharing order dispatching with mean field multi-agent reinforcement learning. In: The World Wide Web Conference, WWW 2019, San Francisco, CA, USA, 13–17 May 2019. pp. 983–994. ACM (2019)
Lillicrap, T.P., et al.: Continuous control with deep reinforcement learning. In: Bengio, Y., LeCun, Y. (eds.) 4th International Conference Track Proceedings on Learning Representations, ICLR 2016, San Juan, Puerto Rico, 2–4 May 2016 (2016)
Mnih, M., et al.: Human-level control through deep reinforcement learning. Nature 518(7540), 529–533 (2015)
Neiat, A.G., Bouguettaya, A., Mistry, S.: Incentive-based crowdsourcing of hotspot services. ACM Trans. Internet Techn. 19(1), 5:1–5:24 (2019)
Pan, L., Cai, Q., Fang, Z., Tang, P., Huang, L.: A deep reinforcement learning framework for rebalancing dockless bike sharing systems. In: The Thirty-Third AAAI Conference on Artificial Intelligence, AAAI 2019, pp. 1393–1400. AAAI Press (2019)
Qin, Z.T., et al.: Ride-hailing order dispatching at DIDI via reinforcement learning. INFORMS J. Appl. Anal. 50(5), 272–286 (2020)
Said, A.B., Erradi, A.: Deep-gap: a deep learning framework for forecasting crowdsourcing supply-demand gap based on imaging time series and residual learning. In: 2019 IEEE International Conference on Cloud Computing Technology and Science (CloudCom), Sydney, Australia, 11–13 December 2019. pp. 279–286. IEEE (2019)
Said, A.B., Erradi, A.: Multiview topological data analysis for crowdsourced service supply-demand gap prediction. In: 16th International Wireless Communications and Mobile Computing Conference, IWCMC 2020, Limassol, Cyprus, 15–19 June 2020. pp. 1818–1823. IEEE (2020)
Singla, A., Santoni, M., Bartók, G., Mukerji, P., Meenen, M., Krause, A.: Incentivizing users for balancing bike sharing systems. In: Proceedings of the Twenty-Ninth AAAI Conference on Artificial Intelligence, 25–30 January 2015, Austin, Texas, USA, pp. 723–729. AAAI Press (2015)
Vaswani, A., et al.: Attention is all you need. In: Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, 4–9 December 2017, Long Beach, CA, USA, pp. 5998–6008 (2017)
Wang, S., Chen, H., Cao, J., Zhang, J., Yu, P.S.: Locally balanced inductive matrix completion for demand-supply inference in stationless bike-sharing systems. IEEE Trans. Knowl. Data Eng. 32(12), 2374–2388 (2020)
Wang, Z., Qin, Z.T., Tang, X., Ye, J., Zhu, H.: Deep reinforcement learning with knowledge transfer for online rides order dispatching. In: IEEE International Conference on Data Mining, ICDM 2018, Singapore, 17–20 November 2018. pp. 617–626. IEEE Computer Society (2018)
Acknowledgements
This work was supported by National Key R &D Program of China (No.2021YFF0900802), Natural Science Foundation of China (No. 91846205) and Natural Science Foundation of Shandong Province (No. ZR2019LZH008).
Author information
Authors and Affiliations
Corresponding authors
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2022 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Li, Z., He, W., Liu, N., Xu, Y., Cui, L., Qi, K. (2022). Balancing Supply and Demand for Mobile Crowdsourcing Services. In: Troya, J., Medjahed, B., Piattini, M., Yao, L., Fernández, P., Ruiz-Cortés, A. (eds) Service-Oriented Computing. ICSOC 2022. Lecture Notes in Computer Science, vol 13740. Springer, Cham. https://doi.org/10.1007/978-3-031-20984-0_20
Download citation
DOI: https://doi.org/10.1007/978-3-031-20984-0_20
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-20983-3
Online ISBN: 978-3-031-20984-0
eBook Packages: Computer ScienceComputer Science (R0)