Abstract
It is reported that task monetary prize is one of the most important motivating factors to attract crowd workers. While using expert-based methods to price Crowdsourcing tasks is a common practice, the challenge of validating the associated prices across different tasks is a constant issue. To address this issue, three different classifications of multiple linear regression, logistic regression, and K-nearest neighbor were compared to find the most accurate predicted price, using a dataset from TopCoder website. The result of comparing chosen algorithms showed that the logistics regression model will provide the highest accuracy of 90% to predict the associated price to tasks and KNN ranked the second with an accuracy of 64% for K = 7. Also, applying PCA wouldn’t lead to any better prediction accuracy as data components are not correlated.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Stol, K.-J., Fitzgerald, B.: Two’s company, three’s a crowd: a case study of crowdsourcing software development. In: The 36th International Conference on Software Engineering (2014)
Faradani, S., Hartmann, B., Ipeirotis, P.G.: What’s the right price? Pricing tasks for finishing on time. In: Proceedings of the Human Computation (2011)
Archak, N.: Money, glory and cheap talk: analyzing strategic behavior of contestants in simultaneous crowdsourcing contests on topcoder.com. In: Proceedings of the 19th International Conference on World Wide Web (WWW 2010), New York, NY, USA, pp. 21–30 (2010)
Wang, L., Yang, Y., Wang, Y.: Do higher incentives lead to better performance?-an exploratory study on software crowdsourcing. In: 2019 ACM/IEEE International Symposium on Empirical Software Engineering and Measurement (ESEM) (2019)
Alelyani, T., Mao, K., Yang, Y.: Context-centric pricing: early pricing models for software crowdsourcing tasks. In: PROMISE: Proceedings of the 13th International Conference on Predictive Models and Data Analytics in Software Engineering, pp. 63–72, November 2017
Yang, Y., Saremi, R.: Award vs. worker behaviors in competitive crowdsourcing tasks. In: ESEM 2015, pp. 1–10 (2015)
Mao, K., Yang, Y., Li, M., Harman, M.: Pricing crowdsourcing-based software development tasks, Piscataway, NJ, USA, pp. 1205–1208 (2013)
Topcoder website. http://www.topcoder.com
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2020 Springer Nature Switzerland AG
About this paper
Cite this paper
Lotfalian Saremi, M., Saremi, R., Martinez-Mejorado, D. (2020). How Much Should I Pay? An Empirical Analysis on Monetary Prize in TopCoder. In: Stephanidis, C., Antona, M. (eds) HCI International 2020 - Posters. HCII 2020. Communications in Computer and Information Science, vol 1226. Springer, Cham. https://doi.org/10.1007/978-3-030-50732-9_27
Download citation
DOI: https://doi.org/10.1007/978-3-030-50732-9_27
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-50731-2
Online ISBN: 978-3-030-50732-9
eBook Packages: Computer ScienceComputer Science (R0)