Abstract
Collective decision-making (CDM) processes – wherein the knowledge of a group of individuals with a common goal must be combined to make optimal decisions – can be formalized within the framework of the deciding with expert advice setting. Traditional approaches to tackle this problem focus on finding appropriate weights for the individuals in the group. In contrast, we propose here meta-CMAB, a meta approach that learns a mapping from expert advice to expected outcomes. In summary, our work reveals that, when trying to make the best choice in a problem with multiple alternatives, meta-CMAB assures that the collective knowledge of experts leads to the best outcome without the need for accurate confidence estimates.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
- 1.
Neither EXP4 (and derivatives) nor meta-CMAB make explicit use of the base contexts to solve the problem.
- 2.
Code available at https://github.com/axelabels/CDM.
References
Reily, K., Finnerty, P.L., Terveen, L.: Two peers are better than one: aggregating peer reviews for computing assignments is surprisingly accurate. In: Proceedings of the ACM 2009 International Conference on Supporting Group Work, pp. 115–124. ACM (2009). https://doi.org/10.1145/1531674.1531692
Robson, N., Rew, D.: Collective wisdom and decision making in surgical oncology. Eur. J. Surg. Oncol. (EJSO) 36(3), 230–236 (2010). https://doi.org/10.1016/j.ejso.2010.01.002
Li, L., Chu, W., Langford, J., Schapire, R.E.: A contextual-bandit approach to personalized news article recommendation. In: Proceedings of the 19th International Conference on World Wide Web, pp. 661–670. ACM (2010)
Aikenhead, G.S.: Collective decision making in the social context of science. Sci. Educ. 69(4), 453–475 (1985). https://doi.org/10.1002/sce.3730690403
Nitzan, S., Paroush, J.: Collective Decision Making: An Economic Outlook. CUP Archive, Cambridge (1985). https://doi.org/10.1016/0167-2681(87)90033-3
Auer, P., Cesa-Bianchi, N., Freund, Y., Schapire, R.: The nonstochastic multiarmed bandit problem. SIAM J. Comput. 32(1), 48–77 (2002). https://doi.org/10.1137/S0097539701398375
McMahan, H.B., Streeter, M.: Tighter bounds for multi-armed bandits with expert advice. In: Proceedings of the 22nd Annual Conference on Learning Theory (COLT) (2009)
Zhou, L.: A survey on contextual multi-armed bandits. CoRR abs/1508.03326 (2015)
Abels, A., Lenaerts, T., Trianni, V., Nowé, A.: How expert confidence can improve collective decision-making in contextual multi-armed bandit problems. In: Nguyen, T.N., Hoang, H.B., Huynh, P.C., Hwang, D., Trawiński, B., Vossen, G. (eds.) ICCCI 2020. LNCS, vol. 11683, pp. 125–138. Springer, Cham (2020)
Marshall, J.A., Brown, G., Radford, A.N.: Individual confidence-weighting and group decision-making. Trends Ecol. Evol. 32(9), 636–645 (2017). https://doi.org/10.1016/j.tree.2017.06.004
Elmachtoub, A.N., McNellis, R., Oh, S., Petrik, M.: A practical method for solving contextual bandit problems using decision trees. CoRR abs/1706.04687 (2017)
Valko, M., Korda, N., Munos, R., Flaounas, I.N., Cristianini, N.: Finite-time analysis of kernelised contextual bandits. CoRR abs/1309.6869 (2013)
Lagae, A., et al.: A survey of procedural noise functions. In: Computer Graphics Forum, vol. 29, pp. 2579–2600. Wiley Online Library (2010). https://doi.org/10.1111/j.1467-8659.2010.01827.x
Thompson, W.R.: On the likelihood that one unknown probability exceeds another in view of the evidence of two samples. Biometrika 25(3/4), 285–294 (1933). https://doi.org/10.2307/2332286
Agrawal, R.: Sample mean based index policies by o (log n) regret for the multi-armed bandit problem. Adv. Appl. Probab. 27(4), 1054–1078 (1995). https://www.jstor.org/stable/1427934
Chu, W., Li, L., Reyzin, L., Schapire, R.: Contextual bandits with linear payoff functions. In: Proceedings of the Fourteenth International Conference on Artificial Intelligence and Statistics, pp. 208–214 (2011)
Agrawal, S., Goyal, N.: Further optimal regret bounds for Thompson sampling. CoRR abs/1209.3353 (2012)
Acknowledgment
This publication benefits from the support of the French Community of Belgium in the context of a FRIA grant, and by the FuturICT2.0 (www.futurict2.eu) project funded by the FLAG-ERA Joint Transnational call (JTC) 2016.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2020 Springer Nature Switzerland AG
About this paper
Cite this paper
Abels, A., Lenaerts, T., Trianni, V., Nowé, A. (2020). Collective Decision-Making as a Contextual Multi-armed Bandit Problem. In: Nguyen, N.T., Hoang, B.H., Huynh, C.P., Hwang, D., Trawiński, B., Vossen, G. (eds) Computational Collective Intelligence. ICCCI 2020. Lecture Notes in Computer Science(), vol 12496. Springer, Cham. https://doi.org/10.1007/978-3-030-63007-2_9
Download citation
DOI: https://doi.org/10.1007/978-3-030-63007-2_9
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-63006-5
Online ISBN: 978-3-030-63007-2
eBook Packages: Computer ScienceComputer Science (R0)