Abstract
The use of artificial intelligence (AI) in hiring entails vast ethical challenges. As such, using an ethical lens to study this phenomenon is to better understand whether and how AI matters in hiring. In this paper, we examine whether ethical perceptions of using AI in the hiring process influence individuals’ trust in the organizations that use it. Building on the organizational trust model and the unified theory of acceptance and use of technology, we explore whether ethical perceptions are shaped by individual differences in performance expectancy and social influence and how they, in turn, impact organizational trust. We collected primary data from over 300 individuals who were either active job seekers or who had recent hiring experience to capture perceptions across the full range of hiring methods. Our findings indicate that performance expectancy, but not social influence, impacts the ethical perceptions of AI in hiring, which in turn influence organizational trust. Additional analyses indicate that these findings vary depending on the type of hiring methods AI is used for, as well as on whether participants are job seekers or individuals with hiring experience. Our study offers theoretical and practical implications for ethics in HRM and informs policy implementation about when and how to use AI in hiring methods, especially as it pertains to acting ethically and trustworthily.
Similar content being viewed by others
Data Availability
The data that support the findings of this study are available from the corresponding author upon reasonable request.
Notes
“We have high-level machine intelligence when machines are able to perform almost all tasks that are economically relevant today better than the median human (today) at each task.” (Zhang and Dafoe, 2019, p. 34).
References
Accenture. (2015). U.S. consumers want more personalized retail experience and control over personal information, Accenture Survey Shows. Retrieved May 3, 2021, from https://newsroom.accenture.com/industries/retail/us-consumers-want-more-personalized-retail-experience-and-control-over-personal-information-accenture-survey-shows.htm
Adell, E., Várhelyi, A., & Nilsson, L. (2018). The definition of acceptance and acceptability. In Driver acceptance of new technology (pp. 11–22). CRC Press.
Anderson, N. (2003). Applicant and recruiter reactions to new technology in selection: A critical review and agenda for future research. International Journal of Selection and Assessment, 11(2–3), 121–136.
Araujo, T., Helberger, N., Kruikemeier, S., & De Vreese, C. H. (2020). In AI we trust? Perceptions about automated decision-making by artificial intelligence. AI & Society, 35(3), 611–623.
Barocas, S., & Selbst, A. D. (2016). Big data’s disparate impact. California Law Review, 104, 671–732.
Bauer, T. N., Truxillo, D. M., Mansfield, L. R., & Erdogan, B. (2012). Contingent workers: Who are they and how can we select them for success?. In The Oxford handbook of personnel assessment and selection.
BBC. (2018). Artificial Intelligence: Morality in the 21st century. Retrieved June 7, 2021, from https://www.bbc.co.uk/programmes/b0bgrw3k
Biswas, M. K., & Suar, D. (2016). Antecedents and consequences of employer branding. Journal of Business Ethics, 136(1), 57–72.
Bloomberg, J. (2018). Don’t Trust Artificial Intelligence? Time to Open the AI Black Box. Forbes. Retrieved May 3, 2021, from https://www.forbes.com/sites/jasonbloomberg/2018/09/16/dont-trust-artificialintelligence-time-to-open-the-ai-black-box/#577a14153b4a
Bozan, K., Parker, K., & Davey, B. (2016). A closer look at the social influence construct in the UTAUT Model: An institutional theory-based approach to investigate health IT adoption patterns of the elderly. In Proceedings of the 49th Hawaii international conference on system sciences (pp. 3105–3114).
Brooksbank, R., Fullerton, S., & Miller, S. (2019). Technology-based marketing strategies through the consumer lens: How might perceptions of ethicality and effectiveness interrelate? International Journal of Technology Marketing, 13(3–4), 428–451.
Brynjolfsson, E., & McAfee, A. (2014). The second machine age: Work, progress, and prosperity in a time of brilliant technologies. WW Norton & Company.
Brynjolfsson, E., Rock, D., & Syverson, C. (2019). A clash of expectations and statistics. In A. Agrawal, J. Gans, & A. Goldfarb (Eds.), Artificial intelligence and the modern productivity paradox (pp. 23–60). University of Chicago Press.
Buolamwini, J., & Gebru, T. (2018). Gender shades: Intersectional accuracy disparities in commercial gender classification. In Conference on fairness, accountability and transparency (pp. 77–91).
Callan, M. J., Kim, H., Gheorghiu, A. I., & Matthews, W. J. (2017). The interrelations between social class, personal relative deprivation, and prosociality. Social Psychological and Personality Science, 8(6), 660–669.
CareerBuilder. (2017, May 18). More than half of HR managers say AI will become a regular part of HR in next 5 years. Retrieved June 15, 2021, from http://press.careerbuilder.com/2017-05-18-More-Than-Half-of-HR-Managers-Say-Artificial-Intelligence-Will-Become-a-Regular-Part-of-HR-in-Next-5-Years
Chamorro-Premuzic, T., Polli, F., & Dattner, B. (2019). Building ethical AI for talent management. Harvard Business Review, 21.
Chattaraman, V., Kwon, W. S., Gilbert, J. E., & Li, Y. (2014). Virtual shopping agents. Journal of Research in Interactive Marketing, 8(2), 144–162.
Choi, E., Bahadori, M. T., Schuetz, A., Stewart, W. F., & Sun, J. (2016). Doctor AI: Predicting clinical events via recurrent neural networks. In Proceedings of the 1st machine learning for healthcare conference (Vol. 56, pp. 301–318).
Clark, B. B., Robert, C., & Hampton, S. A. (2016). The technology effect: How perceptions of technology drive excessive optimism. Journal of Business and Psychology, 31(1), 87–102.
Cordeiro, W. P. (1997). Suggested management responses to ethical issues raised by technological change. Journal of Business Ethics, 16, 1393–1400.
Currall, S. C., & Inkpen, A. C. (2002). A multilevel approach to trust in joint ventures. Journal of International Business Studies, 33(3), 479–495.
Daniels, N. (1979). Wide reflective equilibrium and theory acceptance in ethics. The Journal of Philosophy, 76(5), 256–282.
Dastin, J. (2018). Amazon scraps secret AI recruiting tool that showed bias against women. Reuters. Retrieved June 15, 2021, from https://www.reuters.com/article/us-amazon-com-jobs-automation-insight-idUSKCN1MK08G
Dattner, B., Chamorro-Premuzic, T., Buchband, R., & Schettler, L. (2019). The legal and ethical implications of using AI in hiring. Harvard Business Review, 25.
Davis, F. D. (1989). Perceived usefulness, perceived ease of use, and user acceptance of information technology. MIS Quarterly, 13(3), 319–340.
Davis, F. D., Bagozzi, R. P., & Warshaw, P. R. (1989). User acceptance of computer technology: A comparison of two theoretical models. Management Science, 35(8), 982–1002.
Davis, F. D., Schoorman, F. D., Mayer, R. C., & Tan, H. H. (2000). The trusted general manager and business unit performance: Empirical evidence of a competitive advantage. Strategic Management Journal, 21(5), 563–576.
Du, S. (2021). Reimagining the future of technology: The Social dilemma review. Journal of Business Ethics, 177(1), 213–215.
Fan, X., Oh, S., McNeese, M., Yen, J., Cuevas, H., Strater, L., & Endsley, M. R. (2008). The influence of agent reliability on trust in human–agent collaboration. In Proceedings of the 15th European conference on cognitive ergonomics: The ergonomics of cool interaction (Vol. 369, pp. 1–8).
Fatma, M., & Rahman, Z. (2017). An integrated framework to understand how consumer-perceived ethicality influences consumer hotel brand loyalty. Service Science, 9(2), 136–146.
Ferrario, A., Loi, M., & Viganò, E. (2020). In AI we trust Incrementally: A Multi-layer model of trust to analyze Human-Artificial intelligence interactions. Philosophy & Technology, 33(3), 523–539.
Fishbein, M., & Ajzen, I. (1977). Belief, attitude, intention, and behavior: An introduction to theory and research. Philosophy and Rhetoric, 10(2), 130–132.
Gaudiello, I., Zibetti, E., Lefort, S., Chetouani, M., & Ivaldi, S. (2016). Trust as indicator of robot functional and social acceptance. An experimental study on user conformation to iCub answers. Computers in Human Behavior, 61, 633–655.
Gibney, E. (2016). Google AI algorithm masters ancient game of Go. Nature News, 529(7587), 445.
Gill, H., Boies, K., Finegan, J. E., & McNally, J. (2005). Antecedents of trust: Establishing a boundary condition for the relation between propensity to trust and intention to trust. Journal of Business and Psychology, 19(3), 287–302.
Glikson, E., & Woolley, A. W. (2020). Human trust in artificial intelligence: Review of empirical research. Academy of Management Annals, 14(2), 627–660.
Gonzalez-Garcia, C. G., Meana-Llorian, D., & Lovelle, J. M. C. (2017). A review about smart objects, sensors, and actuators. International Journal of Interactive Multimedia & Artificial Intelligence, 4(3), 7–10.
Grace, K., Salvatier, J., Dafoe, A., Zhang, B., & Evans, O. (2018). When will AI exceed human performance? Evidence from AI experts. Journal of Artificial Intelligence Research, 62, 729–754.
Greenwood, M., & Van Buren III, H. J. (2010). Trust and stakeholder theory: Trustworthiness in the organisation–stakeholder relationship. Journal of Business Ethics, 95(3), 425–438.
Gunz, S., & Thorne, L. (2020). Thematic Symposium: The Impact of Technology on Ethics, Professionalism and Judgement in Accounting. Journal of Business Ethics, 167, 153–155.
Gupta, R., Jain, K., & Jajodia, I. (2021). Determinants of smart speaker adoption intention: Extending the theory of planned behaviour. International Journal of Technology Marketing, 15(2–3), 181–202.
Haenlein, M., Huang, M. H., & Kaplan, A. (2022). Guest Editorial: Business ethics in the era of artificial intelligence. Journal of Business Ethics. https://doi.org/10.1007/s10551-022-05060-x
Hair, J. F., Anderson, R. E., Tatham, R. L., & Black, W. C. (1998). Multivariate data analysis (Fifth edition). Spain Prentice Hall.
Hermann, E. (2021). Leveraging artificial intelligence in marketing for social good—An ethical perspective. Journal of Business Ethics. https://doi.org/10.1007/s10551-021-04843-y
Hinkin, T. R. (1998). A brief tutorial on the development of measures for use in survey questionnaires. Organizational Research Methods, 1(1), 104–121.
Howard, J. (2019). Artificial intelligence: Implications for the future of work. American Journal of Industrial Medicine, 62(11), 917–926.
Hrubes, D., Ajzen, I., & Daigle, J. (2001). Predicting hunting intentions and behavior: An application of the theory of planned behavior. Leisure Sciences, 23(3), 165–178.
Hu, L. T., & Bentler, P. M. (1999). Cutoff criteria for fit indexes in covariance structure analysis: Conventional criteria versus new alternatives. Structural Equation Modeling: A Multidisciplinary Journal, 6(1), 1–55.
Hume, D. (2000). An enquiry concerning human understanding: A critical edition (Vol. 3). Oxford University Press.
Hunkenschroer, A. L., & Luetge, C. (2022). Ethics of AI-enabled recruiting and selection: A review and research agenda. Journal of Business Ethics. https://doi.org/10.1007/s10551-022-05049-6
IBM. (2018a). Power your candidate experience with AI. Retrieved September 23, 2018a, from https://twitter.com/IBMWatsonTalent?lang=en
IBM. (2018b). Bias in AI: How we build fair AI systems and less-biased humans. Retrieved April 29, 2021, from https://www.ibm.com/blogs/policy/bias-in-ai/
Jagger, S., Siala, H., & Sloan, D. (2016). It’s all in the game: A 3D learning model for business ethics. Journal of Business Ethics, 137(2), 383–403.
Jan, P. T., Lu, H. P., & Chou, T. C. (2012). The adoption of e-learning: An institutional theory perspective. Turkish Online Journal of Educational Technology—TOJET, 11(3), 326–343.
Jasanoff, S. (2016). The ethics of invention: Technology and the human future. W. W. Norton & Company.
Johnson, D. G. (2015). Technology with no human responsibility? Journal of Business Ethics, 127(4), 707.
Kaplan, F. (2004). Who is afraid of the humanoid? Investigating cultural differences in the acceptance of robots. International Journal of Humanoid Robotics, 1(3), 465–480.
Keh, H. T., & Xie, Y. (2009). Corporate reputation and customer behavioral intentions: The roles of trust, identification and commitment. Industrial Marketing Management, 38(7), 732–742.
Kijsanayotin, B., Pannarunothai, S., & Speedie, S. M. (2009). Factors influencing health information technology adoption in Thailand’s community health centers: Applying the UTAUT model. International Journal of Medical Informatics, 78(6), 404–416.
Klotz, A. C., da Motta Veiga, S. P., Buckley, M. R., & Gavin, M. B. (2013). The role of trustworthiness in recruitment and selection: A review and guide for future research. Journal of Organizational Behavior, 34(S1), S104–S119.
Knight, W. (2016). Tougher Turing test exposes Chatbots’ stupidity. Retrieved July 8, 2021, from https://www.technologyreview.com/2016/07/14/7797/tougher-turing-test-exposes-chatbots-stupidity/
Lambrecht, A., & Tucker, C. (2019). Algorithmic bias? An empirical study of apparent gender-based discrimination in the display of stem career ads. Management Science, 65(7), 2966–2981.
Laurim, V., Arpaci, S., Prommegger, B., & Krcmar, H. (2021, January). Computer, whom should I hire? Acceptance criteria for artificial intelligence in the recruitment process. In Proceedings of the 54th Hawaii international conference on system sciences (pp. 5495–5504).
Leclercq-Vandelannoitte, A. L. (2017). An ethical perspective on emerging forms of ubiquitous IT-based control. Journal of Business Ethics, 142(1), 139–154.
Lee, M. K. (2018). Understanding perception of algorithmic decisions: Fairness, trust, and emotion in response to algorithmic management. Big Data & Society, 5(1), 2053951718756684.
Leicht-Deobald, U., Busch, T., Schank, C., Weibel, A., Schafheitle, S., Wildhaber, I., & Kasper, G. (2019). The challenges of algorithm-based HR decision-making for personal integrity. Journal of Business Ethics, 160(2), 377–392.
Levy, D. (2009). Love and sex with robots: The evolution of human–robot relationships (p. 352). Harper.
Lewin, K. (1943). Forces behind food habits and methods of change. Bulletin of the National Research Council, 108, 35–65.
Li, P. P., Bai, Y., & Xi, Y. (2012). The contextual antecedents of organizational trust: A multidimensional cross-level analysis. Management and Organization Review, 8(2), 371–396.
Liang, H., Saraf, N., Hu, Q., & Xue, Y. (2007). Assimilation of enterprise systems: The effect of institutional pressures and the mediating role of top management. MIS Quarterly, 31(1), 59–87.
Liao, S. M. (2020). Ethics of artificial intelligence. Oxford University Press.
Lin, C. P. (2010). Modeling corporate citizenship, organizational trust, and work engagement based on attachment theory. Journal of Business Ethics, 94(4), 517–531.
Liu, I. F., Chen, M. C., Sun, Y. S., Wible, D., & Kuo, C. H. (2010). Extending the TAM model to explore the factors that affect intention to use an online learning community. Computers & Education, 54(2), 600–610.
Lockey, S., Gillespie, N., Holm, D., & Someh, I. A. (2021). A review of trust in artificial intelligence: Challenges, vulnerabilities and future directions. In Proceedings of the 54th Hawaii international conference on system sciences (pp. 5463–5472).
Logg, J. M., Minson, J. A., & Moore, D. A. (2019). Algorithm appreciation: People prefer algorithmic to human judgment. Organizational Behavior and Human Decision Processes, 151, 90–103.
Loureiro, S. M., Cavallero, L., & Miranda, F. J. (2018). Fashion brands on retail websites: Customer performance expectancy and e-word-of-mouth. Journal of Retailing and Consumer Services, 41, 131–141.
Madhavan, R., & Grover, R. (1998). From embedded knowledge to embodied knowledge: New product development as knowledge management. Journal of Marketing, 62(4), 1–12.
Margolis, J. D., Grant, A. M., & Molinsky, A. L. (2007). Expanding ethical standards of HRM: Necessary evils and the multiple dimensions of impact. In A. H. Pinnington, R. Macklin, & T. Campbell (Eds.), Human resource management: Ethics and employment (pp. 237–251). Oxford University Press.
Marin, L., Ruiz, S., & Rubio, A. (2009). The role of identity salience in the effects of corporate social responsibility on consumer behavior. Journal of Business Ethics, 84(1), 65–78.
Marreiros, H., Tonin, M., Vlassopoulos, M., & Schraefel, M. C. (2017). Now that you mention it: A survey experiment on information, inattention and online privacy. Journal of Economic Behavior & Organization, 140, 1–17.
Martin, K. (2019). Ethical implications and accountability of algorithms. Journal of Business Ethics, 160, 835–850.
Martin, K. E., & Freeman, R. E. (2004). The separation of technology and ethics in business ethics. Journal of Business Ethics, 53(4), 353–364.
Martin, K., Shilton, K., & Smith, J. (2019). Business and the ethical implications of technology. Journal of Business Ethics, 160, 307–317.
Martin, K. E., & Waldman, A. E. (2022). Are algorithmic decisions legitimate? The effect of process and outcomes on perceptions of legitimacy of AI decisions. Journal of Business Ethics. https://doi.org/10.1007/s10551-021-05032-7
Mayer, R. C., Davis, J. H., & Schoorman, F. D. (1995). An integrative model of organizational trust. Academy of Management Review, 20(3), 709–734.
McCarthy, J. M., Bauer, T. N., Truxillo, D. M., Anderson, N. R., Costa, A. C., & Ahmed, S. M. (2017). Applicant perspectives during selection: A review addressing So what? What’s new? And where to next? Journal of Management, 43(6), 1693–1725.
Meyer, D. (2018). Amazon reportedly killed an AI recruitment system because it couldn’t stop the tool from discriminating against women. Fortune. Retrieved May 3, 2021, from http://fortune.com/2018/10/10/amazon-ai-recruitment-bias-women-sexist/
Morgan, R. M., & Hunt, S. D. (1994). The commitment-trust theory of relationship marketing. Journal of Marketing, 58(3), 20–38.
Müller-Abdelrazeq, S. L., Schönefeld, K., Haberstroh, M., & Hees, F. (2019). Interacting with collaborative robots—a study on attitudes and acceptance in industrial contexts. In O. Korn (Ed.), Social robots: Technological, societal and ethical aspects of human–robot interaction (pp. 101–117). Springer.
Munoko, I., Brown-Liburd, H. L., & Vasarhelyi, M. (2020). The ethical implications of using artificial intelligence in auditing. Journal of Business Ethics, 167, 209–234.
Nawaz, N. (2019). How far have we come with the study of artificial intelligence for recruitment process. International Journal of Scientific & Technology Research, 8(07), 488–493.
Nikolaou, I., Georgiou, K., Bauer, T. N., & Truxillo, D. M. (2019). Applicant reactions in employee recruitment and selection: The role of technology. In R. N. Landers (Ed.), The Cambridge handbook of tech and employee behavior (pp. 100–130). Cambridge University Press.
Noble, S. U. (2018). Algorithms of oppression: How search engines reinforce racism. New York University Press.
North-Samardzic, A. (2020). Biometric technology and ethics: Beyond security applications. Journal of Business Ethics, 167(3), 433–450.
O’Neil, C. (2016). Weapons of math destruction: How big data increases inequality and threatens democracy. Broadway Books.
OECD. (2019). Artificial Intelligence on Society (Vol. 58(3), pp. 377–400). Retrieved from https://www.oecd-ilibrary.org/.ors
OED. (2021). Oxford University Press. OED. Retrieved April 29, 2021, from www.oxfordreference.com
Oshlyansky, L., Cairns, P., & Thimbleby, H. (2007). Validating the unified theory of acceptance and use of technology (UTAUT) tool cross-culturally. In Proceedings of the 21st British HCI group annual conference. University of Lancaster (Vol. 21, pp. 1–4).
Palan, S., & Schitter, C. (2018). Prolific.ac—A subject pool for online experiments. Journal of Behavioral and Experimental Finance, 17, 22–27.
Parmigiani, A., & Mitchell, W. (2005). How buyers shape supplier performance: Can governance skills substitute for technical expertise in managing out-sourcing relationships? Academy of Management Proceedings, 2005(1), C1–C6.
Parry, K. W., Cohen, M., & Bhattacharya, S. (2016). Rise of the machines: A critical consideration of automated leadership decision making in organizations. Group & Organization Management, 41(5), 571–594.
Pasquale, F. (2015). The Black Box Society: The Secret Algorithms that Control Money and Information. Harvard University Press.
Peck, D. (2013). They’re watching you at work. The Atlantic, 312(5), 72–84.
Peer, E., Brandimarte, L., Samat, S., & Acquisti, A. (2017). Beyond the Turk: Alternative platforms for crowdsourcing behavioral research. Journal of Experimental Social Psychology, 70, 153–163.
Pirson, M., & Malhotra, D. (2011). Foundations of organizational trust: What matters to different stakeholders? Organization Science, 22(4), 1087–1104.
Pirson, M., Martin, K., & Parmar, B. (2019). Public trust in business and its determinants. Business & Society, 58(1), 132–166.
Polli, F. (2019). Using AI to eliminate bias from hiring. Harvard Business Review, 29.
Prpic, N. (2020). The AI recruitment evolution—from Amazon’s biased algorithm to contextual understanding. Retrieved May 3, 2021, from https://www.talentlyft.com/en/blog/article/414/the-ai-recruitment-evolution-from-amazons-biased-algorithm-to-contextual-understanding
Pulakos, E. D. (2005). Selection assessment methods. United stated of America: Society for Human Resource Management (SHRM) Foundation.
Rąb-Kettler, K., & Lehnervp, B. (2019). Recruitment in the times of machine learning. Management Systems in Production Engineering, 27, 105–109.
Raghavan, M., Barocas, S., Kleinberg, J., & Levy, K. (2020). Mitigating bias in algorithmic hiring: Evaluating claims and practices. In Proceedings of the 2020 conference on fairness, accountability, and transparency (pp. 469–481).
Ramchurn, S. D., Wu, F., Jiang, W., Fischer, J. E., Reece, S., Roberts, S., Rodden, T., Greenhalgh, C., & Jennings, N. R. (2016). Human–agent collaboration for disaster response. Autonomous Agents and Multi-Agent Systems, 30(1), 82–111.
Rawls, J. (2001). Justice as fairness: A restatement. Harvard University Press.
Robinson, L., Gibson, G., Kingston, A., Newton, L., Pritchard, G., Finch, T., & Brittain, K. (2013). Assistive technologies in caring for the oldest old: A review of current practice and future directions. Aging and Health, 9(4), 365–375.
Rogers, E. M. (1995). Diffusion of innovations (4th ed.). Free Press.
Ryan, A. M., & Ployhart, R. E. (2000). Applicants’ perceptions of selection procedures and decisions: A critical review and agenda for the future. Journal of Management, 26(3), 565–606.
Sanchez-Monedero, J., Dencik, L., & Edwards, L. (2020, January). What does it mean to ‘solve’ the problem of discrimination in hiring? In Proceedings of the 2020 conference on fairness, accountability, and transparency (pp. 458–468).
Schoorman, F. D., Mayer, R. C., & Davis, J. H. (2007). An integrative model of organizational trust: Past, present, and future. Academy of Management Review, 32(2), 344–354.
Schwoerer, C. E., May, D. R., Hollensbe, E. C., & Mencl, J. (2005). General and specific self-efficacy in the context of a training intervention to enhance performance expectancy. Human Resource Development Quarterly, 16(1), 111–129.
Sheeran, P., & Webb, T. L. (2016). The intention–behavior gap. Social and Personality Psychology Compass, 10(9), 503–518.
Sheppard, B. H., & Sherman, D. M. (1998). The grammars of trust: A model and general implications. Academy of Management Review, 23(3), 422–437.
Shilton, K., Koepfler, J. A., & Fleischmann, K. R. (2013). Charting sociotechnical dimensions of values for design research. The Information Society, 29(5), 259–271.
Siau, K., & Wang, W. (2018). Building trust in artificial intelligence, machine learning, and robotics. Cutter Business Technology Journal, 31(2), 47–53.
Singh, J. J., Iglesias, O., & Batista-Foguet, J. M. (2012). Does having an ethical brand matter? The influence of consumer perceived ethicality on trust, affect and loyalty. Journal of Business Ethics, 111(4), 541–549.
Sohn, K., & Kwon, O. (2020). Technology acceptance theories and factors influencing artificial Intelligence-based intelligent products. Telematics and Informatics, 47, 101324.
Speicher, T., Heidari, H., Grgic-Hlaca, N., Gummadi, K. P., Singla, A., Weller, A., & Zafar, M. B. (2018, July). A unified approach to quantifying algorithmic unfairness: Measuring individual &group unfairness via inequality indices. In Proceedings of the 24th ACM SIGKDD international conference on knowledge discovery & data mining (pp. 2239–2248).
StataCorp. (2019). Stata 16 Base Reference Manual. College Station, TX: Stata Press.
Surowiecki, J. (2005). The wisdom of crowds. Anchor.
Tambe, P., Cappelli, P., & Yakubovich, V. (2019). Artificial intelligence in human resources management: Challenges and a path forward. California Management Review, 61(4), 15–42.
Teo, H. H., Wei, K. K., & Benbasat, I. (2003). Predicting intention to adopt interorganizational linkages: An institutional perspective. MIS Quarterly, 27(1), 19–49.
Thompson, R. L., Higgins, C. A., & Howell, J. M. (1991). Personal computing: Toward a conceptual model of utilization. MIS Quarterly, 15(1), 124–143.
Tilcsik, A. (2021). Statistical discrimination and the rationalization of stereotypes. American Sociological Review, 86(1), 93–122.
Turkle, S. (Ed.). (2011). The inner history of devices. MIT Press.
Upadhyay, A. K., & Khandelwal, K. (2018). Applying artificial intelligence: Implications for recruitment. Strategic HR Review, 17(5), 255–258.
Van de Poel, I. (2016). An ethical framework for evaluating experimental technology. Science and Engineering Ethics, 22(3), 667–686.
van den Broek, E., Sergeeva, A., & Huysman, M. (2019). Hiring algorithms: an ethnography of fairness in practice. In ICIS Proceedings, 6. https://aisel.aisnet.org/icis2019/future_of_work/future_work/6
Venkatesh, V., Morris, M. G., Davis, G. B., & Davis, F. D. (2003). User acceptance of information technology: Toward a unified view. MIS Quarterly, 27(3), 425–478.
Venkatesh, V., Thong, J. Y., & Xu, X. (2012). Consumer acceptance and use of information technology: Extending the unified theory of acceptance and use of technology. MIS Quarterly, 36(1), 157–178.
Venkatesh, V., Thong, J. Y., & Xu, X. (2016). Unified theory of acceptance and use of technology: A synthesis and the road ahead. Journal of the Association for Information Systems, 17(5), 328–376.
Weber, J., & Gillespie, J. (1998). Differences in ethical beliefs, intentions, and behaviors: The role of beliefs and intentions in ethics research revisited. Business & Society, 37(4), 447–467.
Why, M. (2018). 4 reasons why an automated hiring process will help your company. In Select international, a PSI business (Vol. 2018). Select International.
Wright, S. A., & Schultz, A. E. (2018). The rising tide of artificial intelligence and business automation: Developing an ethical framework. Business Horizons, 61(6), 823–832.
Yampolskiy, R. V. (2019). Predicting future AI failures from historic examples. Foresight, 21(1), 138–152.
Zaheer, A., McEvily, B., & Perrone, V. (1998). Does trust matter? Exploring the effects of interorganizational and interpersonal trust on performance. Organization Science, 9(2), 141–159.
Zhang, B., & Dafoe, A. (2019). Artificial intelligence: American attitudes and trends. SSRN 3312874.
Zhang, C., & Lu, Y. (2021). Study on artificial intelligence: The state of the art and future prospects. Journal of Industrial Information Integration, 23, 100224.
Author information
Authors and Affiliations
Corresponding authors
Ethics declarations
Conflict of interest
The authors declare that they have no conflict of interest.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Appendix
Appendix
See Fig. 2.
Rights and permissions
About this article
Cite this article
Figueroa-Armijos, M., Clark, B.B. & da Motta Veiga, S.P. Ethical Perceptions of AI in Hiring and Organizational Trust: The Role of Performance Expectancy and Social Influence. J Bus Ethics 186, 179–197 (2023). https://doi.org/10.1007/s10551-022-05166-2
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10551-022-05166-2