‘If You Agree with Me, Do I Trust You?’: An Examination of Human-Agent Trust from a Psychological Perspective

  • Hsiao-Ying Huang
  • Michael Twidale
  • Masooda BashirEmail author
Conference paper
Part of the Advances in Intelligent Systems and Computing book series (AISC, volume 1038)


Applications of automated agent systems in daily life have changed the role of human operators from a controller to a teammate. However, this ‘teammate’ relationship between humans and agents raises an important but challenging question: how do humans develop trust when interacting with automated agents that are human-like? In this study, a two-phase online experiment was conducted to examine the effect of attitudinal congruence and individual personalities on users’ trust toward an anthropomorphic agent. Our results suggest that the degree of an agent’s response congruence had no significant impacts on users’ trust toward the agent. In terms of individual personalities, we found one personality trait that has significant impact on users’ formation of human-agent trust. Although our data does not support the effect of attitudinal congruence on human-agent trust formation, this study provides the essential empirical evidence that benefits future research in this field. More importantly, in this paper we address the unusual challenges in our experimental design and what our null results imply about the formation of human-agent trust. This study not only sheds light on trust formation in human-agent collaboration but also provides insight for the future design of automated agent systems.


Human-agent interaction Decision-aids automation Trust formation Personality traits Social psychology 


  1. 1.
    Bailey, N.R., Scerbo, M.W.: Automation-induced complacency for monitoring highly reliable systems: the role of task complexity, system experience, and operator trust. Theoret. Issues Ergon. Sci. 8(4), 321–348 (2007)CrossRefGoogle Scholar
  2. 2.
    Behrenbruch, K., Söllner, M., Leimeister, J.M., Schmidt, L.: Understanding diversity–the impact of personality on technology acceptance. In: IFIP Conference on Human-Computer Interaction, pp. 306–313. Springer, Heidelberg, September 2013CrossRefGoogle Scholar
  3. 3.
    Biros, D.P., Fields, G., Gunsch, G.: The effect of external safeguards on human-information system trust in an information warfare environment. In: 2003 Proceedings of the 36th Annual Hawaii International Conference on System Sciences, p. 10. IEEE, January 2003Google Scholar
  4. 4.
    Biros, D.P., Daly, M., Gunsch, G.: The influence of task load and automation trust on deception detection. Group Decis. Negot. 13(2), 173–189 (2004)CrossRefGoogle Scholar
  5. 5.
    Huang, H.Y., Bashir, M.: Personal influences on dynamic trust formation in human-agent interaction. In: Proceedings of the 5th International Conference on Human Agent Interaction, pp. 233–243. ACM, October 2017Google Scholar
  6. 6.
    Cahour, B., Forzy, J.F.: Does projection into use improve trust and exploration? An example with a cruise control system. Safety Sci. 47(9), 1260–1270 (2009)CrossRefGoogle Scholar
  7. 7.
    Chavaillaz, A., Wastell, D., Sauer, J.: System reliability, performance and trust in adaptable automation. Appl. Ergon. 52, 333–342 (2016)CrossRefGoogle Scholar
  8. 8.
    Klien, G., Woods, D.D., Bradshaw, J.M., Hoffman, R.R., Feltovich, P.J.: Ten challenges for making automation a team player in joint human-agent activity. IEEE Intell. Syst. 19(6), 91–95 (2004)CrossRefGoogle Scholar
  9. 9.
    Cummings, M.L., Clare, A., Hart, C.: The role of human-automation consensus in multiple unmanned vehicle scheduling. Hum. Factors 52(1), 17–27 (2010)CrossRefGoogle Scholar
  10. 10.
    de Visser, E., Parasuraman, R.: Adaptive aiding of human-robot teaming: Effects of imperfect automation on performance, trust, and workload. J. Cogn. Eng. Decis. Making 5(2), 209–231 (2011)CrossRefGoogle Scholar
  11. 11.
    Donmez, B., Boyle, L.N., Lee, J.D., McGehee, D.V.: Drivers’ attitudes toward imperfect distraction mitigation strategies. Transp. Res. Part F: Traffic Psychol. Behav. 9(6), 387–398 (2006)CrossRefGoogle Scholar
  12. 12.
    Dzindolet, M.T., Peterson, S.A., Pomranky, R.A., Pierce, L.G., Beck, H.P.: The role of trust in automation reliance. Int. J. Hum.-Comput. Stud. 58(6), 697–718 (2003)CrossRefGoogle Scholar
  13. 13.
    Merritt, S.M.: Affective processes in human–automation interactions. Hum. Factors 53(4), 356–370 (2011)CrossRefGoogle Scholar
  14. 14.
    Epley, N., Waytz, A., Cacioppo, J.T.: On seeing human: a three-factor theory of anthropomorphism. Psychol. Rev. 114(4), 864 (2007)CrossRefGoogle Scholar
  15. 15.
    Hancock, P.A., Billings, D.R., Schaefer, K.E., Chen, J.Y., De Visser, E.J., Parasuraman, R.: A meta-analysis of factors affecting trust in human-robot interaction. Hum. Factors 53(5), 517–527 (2011)CrossRefGoogle Scholar
  16. 16.
    Ho, G., Wheatley, D., Scialfa, C.T.: Age differences in trust and reliance of a medication management system. Interact. Comput. 17(6), 690–710 (2005)CrossRefGoogle Scholar
  17. 17.
    Wetzel, J.M.: Driver trust, annoyance, and compliance for an automated calendar system. Doctoral dissertation. Retrieved from ProQuest Dissertations and Theses (2006)Google Scholar
  18. 18.
    Hoff, K.A., Bashir, M.: Trust in automation: integrating empirical evidence on factors that influence trust. Hum. Factors 57(3), 407–434 (2015)CrossRefGoogle Scholar
  19. 19.
    Dadashi, N., Stedmon, A.W., Pridmore, T.P.: Semi-automated CCTV surveillance: the effects of system confidence, system accuracy and task complexity on operator vigilance, reliance and workload. Appl. Ergon. 44(5), 730–738 (2013)CrossRefGoogle Scholar
  20. 20.
    Singh, I.L., Molloy, R., Parasuraman, R.: Automation-induced complacency: development of the complacency-potential rating scale. Int. J. Aviat. Psychol. 3(2), 111–122 (1993)CrossRefGoogle Scholar
  21. 21.
    Kircher, K., Thorslund, B.: Effects of road surface appearance and low friction warning systems on driver behaviour and confidence in the warning system. Ergonomics 52(2), 165–176 (2009)CrossRefGoogle Scholar
  22. 22.
    Dzindolet, M.T., Pierce, L.G., Beck, H.P., Dawe, L.A.: The perceived utility of human and automated aids in a visual detection task. Hum. Factors 44(1), 79–94 (2002)CrossRefGoogle Scholar
  23. 23.
    Kohn, A.: Brain science: The forgetting curve–the dirty secret of corporate training (2016). Accessed 7 Sept 2016
  24. 24.
    Costa Jr., P.T., McCrae, R.R.: Four ways five factors are basic. Pers. Individ. Differ. 13(6), 653–665 (1992)CrossRefGoogle Scholar
  25. 25.
    Lee, J.D., See, K.A.: Trust in automation: designing for appropriate reliance. Hum. Factors 46(1), 50–80 (2004)CrossRefGoogle Scholar
  26. 26.
    Lee, M.N., Lee, J.D.: The influence of distraction and driving context on driver response to imperfect collision warning systems. Ergonomics 50(8), 1264–1286 (2007)CrossRefGoogle Scholar
  27. 27.
    Looije, R., Neerincx, M.A., Cnossen, F.: Persuasive robotic assistant for health self-management of older adults: design and evaluation of social behaviors. Int. J. Hum.-Comput. Stud. 68(6), 386–397 (2010)CrossRefGoogle Scholar
  28. 28.
    Seong, Y., Bisantz, A.M.: The impact of cognitive feedback on judgment performance and trust with decision aids. Int. J. Ind. Ergon. 38(7–8), 608–625 (2008)CrossRefGoogle Scholar
  29. 29.
    Vlasic, B., Boudette, N.E.: As US investigates fatal Tesla crash, company defends Autopilot system. The New York Times (2016). Accessed 19 September 2017
  30. 30.
    Mayer, R.C., Davis, J.H.: The effect of the performance appraisal system on trust for management: a field quasi-experiment. J. Appl. Psychol. 84(1), 123 (1999)CrossRefGoogle Scholar
  31. 31.
    Jennings, N.R., Moreau, L., Nicholson, D., Ramchurn, S., Roberts, S., Rodden, T., Rogers, A.: Human-agent collectives. Commun. ACM 57(12), 80–88 (2014)CrossRefGoogle Scholar
  32. 32.
    Madhavan, P., Wiegmann, D.A.: Similarities and differences between human–human and human–automation trust: an integrative review. Theoret. Issues Ergon. Sci. 8(4), 277–301 (2007)CrossRefGoogle Scholar
  33. 33.
    Merritt, S.M., Ilgen, D.R.: Not all trust is created equal: dispositional and history-based trust in human-automation interactions. Hum. Factors 50(2), 194–210 (2008)CrossRefGoogle Scholar
  34. 34.
    Miramontes, A., Tesoro, A., Trujillo, Y., Barraza, E., Keeler, J., Boudreau, A., Strybel, T.Z., Vu, K.P.L.: Training student air traffic controllers to trust automation. Procedia Manuf. 3, 3005–3010 (2015)CrossRefGoogle Scholar
  35. 35.
    Szalma, J.L., Taylor, G.S.: Individual differences in response to automation: the five factor model of personality. J. Exp. Psychol.: Appl. 17(2), 71 (2011)Google Scholar
  36. 36.
    Mosier, K.L., Skitka, L.J., Korte, K.J.: Cognitive and social psychological issues in flight crew/automation interaction. In: Human Performance in Automated Systems: Current Research and Trends, pp. 191–197 (1994)Google Scholar
  37. 37.
    Nass, C., Moon, Y., Carney, P.: Are people polite to computers? Responses to computer‐based interviewing systems 1. J. Appl. Soc. Psychol. 29(5), 1093–1109 (1999)CrossRefGoogle Scholar
  38. 38.
    Nass, C., Steuer, J., Tauber, E.R.: Computers are social actors. In: Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, pp. 72–78. ACM (1994)Google Scholar
  39. 39.
    Nomura, T., Kanda, T., Suzuki, T., Kato, K.: Prediction of human behavior in human–robot interaction using psychological scales for anxiety and negative attitudes toward robots. IEEE Trans. Robot. 24(2), 442–451 (2008)CrossRefGoogle Scholar
  40. 40.
    Oduor, K.F., Campbell, C.S.: Deciding when to trust automation in a policy-based city management game: policity. In: Proceedings of the 2007 Symposium on Computer Human Interaction for the Management of Information Technology, p. 2. ACM, March 2007Google Scholar
  41. 41.
    Pak, R., Fink, N., Price, M., Bass, B., Sturre, L.: Decision support aids with anthropomorphic characteristics influence trust and performance in younger and older adults. Ergonomics 55(9), 1059–1072 (2012)CrossRefGoogle Scholar
  42. 42.
    Pak, R., Rovira, E., McLaughlin, A.C., Baldwin, N.: Does the domain of technology impact user trust? Investigating trust in automation across different consumer-oriented domains in young adults, military, and older adults. Theoret. Issues Ergon. Sci. 18(3), 199–220 (2017)CrossRefGoogle Scholar
  43. 43.
    Parasuraman, R., Miller, C.A.: Trust and etiquette in high-criticality automated systems. Commun. ACM 47(4), 51–55 (2004)CrossRefGoogle Scholar
  44. 44.
    Tung, F.W.: Influence of gender and age on the attitudes of children towards humanoid robots. In: International Conference on Human-Computer Interaction, pp. 637–646. Springer, Heidelberg, July 2011Google Scholar
  45. 45.
    Pearson, C.J., Welk, A.K., Boettcher, W.A., Mayer, R.C., Streck, S., Simons-Rudolph, J.M., Mayhorn, C.B.: Differences in trust between human and automated decision aids. In: Proceedings of the Symposium and Bootcamp on the Science of Security, pp. 95–98. ACM, April 2016Google Scholar
  46. 46.
    Reeves, B., Nass, C.I.: The Media Equation: How People Treat Computers, Television, and New Media Like Real People and Places. Cambridge University Press, Cambridge (1996)Google Scholar
  47. 47.
    Rempel, J.K., Holmes, J.G., Zanna, M.P.: Trust in close relationships. J. Pers. Soc. Psychol. 49(1), 95 (1985)CrossRefGoogle Scholar
  48. 48.
    Ross, J.M., Szalma, J.L., Hancock, P.A., Barnett, J.S., Taylor, G.: The effect of automation reliability on user automation trust and reliance in a search-and-rescue scenario. In: Proceedings of the Human Factors and Ergonomics Society Annual Meeting, vol. 52, no. 19, pp. 1340–1344. Sage, Los Angeles, September 2008CrossRefGoogle Scholar
  49. 49.
    Salem, M., Lakatos, G., Amirabdollahian, F., Dautenhahn, K.: Would you trust a (faulty) robot?: effects of error, task type and personality on human-robot cooperation and trust. In: Proceedings of the Tenth Annual ACM/IEEE International Conference on Human-Robot Interaction, pp. 141–148. ACM, March 2015Google Scholar
  50. 50.
    Sanchez, J., Fisk, A.D., Rogers, W.A.: Reliability and age-related effects on trust and reliance of a decision support aid. In: Proceedings of the Human Factors and Ergonomics Society Annual Meeting, vol. 48, no. 3, pp. 586–589. Sage, Los Angeles, September 2004CrossRefGoogle Scholar
  51. 51.
    Sanchez, J., Rogers, W.A., Fisk, A.D., Rovira, E.: Understanding reliance on automation: effects of error type, error distribution, age and experience. Theoret. Issues Ergon. Sci. 15(2), 134–160 (2014)CrossRefGoogle Scholar
  52. 52.
    Schaefer, K.E., Chen, J.Y., Szalma, J.L., Hancock, P.A.: A meta-analysis of factors influencing the development of trust in automation: implications for understanding autonomy in future systems. Hum. Factors 58(3), 377–400 (2016)CrossRefGoogle Scholar
  53. 53.
    Genesereth, M.R.: Software Agents Logic Group. Computer Science, Department Stanford University (1994)Google Scholar
  54. 54.
    Hommel, B., Colzato, L.S.: Interpersonal trust: an event-based account. Front. Psychol. 6, 1399 (2015)Google Scholar
  55. 55.
    Sitkin, S.B., Roth, N.L.: Explaining the limited effectiveness of legalistic “remedies” for trust/distrust. Organ. Sci. 4(3), 367–392 (1993)CrossRefGoogle Scholar
  56. 56.
    Sharples, S., Stedmon, A., Cox, G., Nicholls, A., Shuttleworth, T., Wilson, J.: Flightdeck and air traffic control collaboration evaluation (FACE): evaluating aviation communication in the laboratory and field. Appl. Ergon. 38(4), 399–407 (2007)CrossRefGoogle Scholar
  57. 57.
    Spain, R.D., Bliss, J.P.: The effect of sonification display pulse rate and reliability on operator trust and perceived workload during a simulated patient monitoring task. Ergonomics 51(9), 1320–1337 (2008)CrossRefGoogle Scholar
  58. 58.
    Lee, E.J.: Flattery may get computers somewhere, sometimes: the moderating role of output modality, computer gender, and user gender. Int. J. Hum.-Comput. Stud. 66(11), 789–800 (2008)CrossRefGoogle Scholar
  59. 59.
    Cheshire, C.: Online trust, trustworthiness, or assurance? Daedalus 140(4), 49–58 (2011)CrossRefGoogle Scholar
  60. 60.
    Bates, D., Maechler, M., Bolker, B., Walker, S., Christensen, R.H.B., Singmann, H., Bolker, M.B.: Package ‘lme4, p. 12. Vienna, Austria: R Foundation for Statistical Computing (2014)Google Scholar
  61. 61.
    de Visser, E.J., Krueger, F., McKnight, P., Scheid, S., Smith, M., Chalk, S., Parasuraman, R.: The world is not enough: trust in cognitive agents. In: Proceedings of the Human Factors and Ergonomics Society Annual Meeting, vol. 56, no. 1, pp. 263–267. Sage, Los Angeles, September 2012Google Scholar
  62. 62.
    Griffitt, W., Veitch, R.: Preacquaintance attitude similarity and attraction revisited: ten days in a fall-out shelter. Sociometry 163–173 (1974)CrossRefGoogle Scholar
  63. 63.
    Kandel, D.B.: Homophily, selection, and socialization in adolescent friendships. Am. J. Sociol. 84(2), 427–436 (1978)CrossRefGoogle Scholar
  64. 64.
    Lee, E.J., Nass, C.: Experimental tests of normative group influence and representation effects in computer-mediated communication: when interacting via computers differs from interacting with computers. Hum. Commun. Res. 28(3), 349–381 (2002)Google Scholar
  65. 65.
    Bickmore, T.W., Picard, R.W.: Establishing and maintaining long-term human-computer relationships. ACM Trans. Comput.-Hum. Interact. (TOCHI) 12(2), 293–327 (2005)CrossRefGoogle Scholar
  66. 66.
    Gong, L.: How social is social responses to computers? The function of the degree of anthropomorphism in computer representations. Comput. Hum. Behav. 24(4), 1494–1509 (2008)CrossRefGoogle Scholar
  67. 67.
    Merritt, S.M., Unnerstall, J.L., Lee, D., Huber, K.: Measuring individual differences in the perfect automation schema. Hum. Factors 57(5), 740–753 (2015)CrossRefGoogle Scholar

Copyright information

© Springer Nature Switzerland AG 2020

Authors and Affiliations

  • Hsiao-Ying Huang
    • 1
  • Michael Twidale
    • 2
  • Masooda Bashir
    • 2
    Email author
  1. 1.Illinois Informatics InstituteUrbanaUSA
  2. 2.School of Information SciencesChampaignUSA

Personalised recommendations