Reference: | [1] Adadi, A., and Berrada, M. Peeking inside the black-box: a survey on explainable artificial intelligence (xai). IEEE access 6 (2018), 52138–52160. [2] Akgun,M.,Cagiltay,K.,andZeyrek,D.Theeffectofapologeticerrormessagesand mood states on computer users’self-appraisal of performance. Journal of Pragmat- ics 42, 9 (2010), 2430–2448. [3] Allen, V. L., and Vliert, E. v. d. A role theoretical perspective on transitional pro- cesses. In Role transitions. Springer, 1984, pp. 3–18. [4] Aronson, E. Readings about the social animal. Macmillan, 2003. [5] Bansal,G.,Nushi,B.,Kamar,E.,Lasecki,W.S.,Weld,D.S.,andHorvitz,E.Beyond accuracy: The role of mental models in human-ai team performance. In Proceedings of the AAAI Conference on Human Computation and Crowdsourcing (2019), vol. 7, pp. 2–11. [6] Basow, S. A. Gender: Stereotypes and roles. Thomson Brooks/Cole Publishing Co, 1992. [7] Benbasat, I., and Wang, W. Trust in and adoption of online recommendation agents. Journal of the association for information systems 6, 3 (2005), 4. [8] Berger, J., Cohen, B. P., and Zelditch Jr, M. Status characteristics and social interaction. American Sociological Review (1972), 241–255. [9] Biddle, B. J. Recent developments in role theory. Annual review of sociology 12, 1 (1986), 67–92. [10] Billings, D. R., Schaefer, K. E., Chen, J. Y., and Hancock, P. A. Human-robot inter- action: developing trust in robots. In Proceedings of the seventh annual ACM/IEEE international conference on Human-Robot Interaction (2012), pp. 109–110. [11] Blut, M., Wang, C., Wünderlich, N. V., and Brock, C. Understanding anthropomorphism in service provision: a meta-analysis of physical robots, chatbots, and other ai. Journal of the Academy of Marketing Science (2021), 1–27. [12] Boon, S. D., and Holmes, J. G. The dynamics of interpersonal trust: Resolving uncertainty in the face of risk. Cooperation and prosocial behavior (1991), 190– 211. [13] Cao, L. Ai in finance: Challenges, techniques, and opportunities. ACM Computing Surveys (CSUR) 55, 3 (2022), 1–38. [14] Chen, J. Y., and Barnes, M. J. Human-agent teaming for multirobot control: A review of human factors issues. IEEE Transactions on Human-Machine Systems 44, 1 (2014), 13–29. [15] Coeckelbergh, M. Can we trust robots? Ethics and information technology 14, 1 (2012), 53–60. [16] Cofta, P. Distrust. In Proceedings of the 8th international conference on Electronic commerce: The new e-commerce: innovations for conquering current barriers, obstacles and limitations to conducting successful business on the internet (2006), pp. 250–258. [17] Daronnat, S., Azzopardi, L., Halvey, M., and Dubiel, M. Impact of agent reliability and predictability on trust in real time human-agent collaboration. In Proceedings of the 8th International Conference on Human-Agent Interaction (2020), pp. 131–139. [18] Dautenhahn, K., Woods, S., Kaouri, C., Walters, M. L., Koay, K. L., and Werry, I. What is a robot companion-friend, assistant or butler? In 2005 IEEE/RSJ international conference on intelligent robots and systems (2005), IEEE, pp. 1192–1197. [19] De Visser, E. J., Monfort, S. S., McKendrick, R., Smith, M. A., McKnight, P. E., Krueger, F., and Parasuraman, R. Almost human: Anthropomorphism increases trust resilience in cognitive agents. Journal of Experimental Psychology: Applied 22, 3 (2016), 331. [20] Deutsch, M. The resolution of conflict: Constructive and destructive processes. Yale University Press, 1973. [21] Evans, N. J., and Jarvis, P. A. The group attitude scale: A measure of attraction to group. Small group behavior 17, 2 (1986), 203–216. [22] Fan, X., Oh, S., McNeese, M., Yen, J., Cuevas, H., Strater, L., and Endsley, M. R. The influence of agent reliability on trust in human-agent collaboration. In Proceedings of the 15th European conference on Cognitive ergonomics: the ergonomics of cool interaction (2008), pp. 1–8. [23] Freedy, A., DeVisser, E., Weltman, G., and Coeyman, N. Measurement of trust in human-robot collaboration. In 2007 international symposium on collaborative technologies and systems (2007), IEEE, pp. 106–114. [24] Glikson, E., and Woolley, A. W. Human trust in artificial intelligence: Review of empirical research. Academy of Management Annals 14, 2 (2020), 627–660. [25] Greenberg, J. Managing behavior in organizations. [26] Haleem, A., Javaid, M., and Khan, I. H. Current status and applications of artificial intelligence (ai) in medical field: An overview. Current Medicine Research and Practice 9, 6 (2019), 231–237. [27] Hamilton, V. L. Who is responsible? toward a social psychology of responsibility attribution. Social Psychology (1978), 316–328. [28] Hamilton,V.L.,andHagiwara,S.Roles, responsibility, and accounts across cultures. International Journal of Psychology 27, 2 (1992), 157–179. [29] Hancock, P. A., Billings, D. R., Schaefer, K. E., Chen, J. Y., De Visser, E. J., and Parasuraman, R. A meta-analysis of factors affecting trust in human-robot interaction. Human factors 53, 5 (2011), 517–527. [30] Hinds, P. J., Roberts, T. L., and Jones, H. Whose job is it anyway? a study of human-robot interaction in a collaborative task. Human-Computer Interaction 19, 1-2 (2004), 151–181. [31] Hoff, K. A., and Bashir, M. Trust in automation: Integrating empirical evidence on factors that influence trust. Human factors 57, 3 (2015), 407–434. [32] Ignat, C.-L., Dang, Q.-V., and Shalin, V. L. The influence of trust score on cooperative behavior. ACM Transactions on Internet Technology (TOIT) 19, 4 (2019), 1–22. [33] Jacovi, A., Marasović, A., Miller, T., and Goldberg, Y. Formalizing trust in artificial intelligence: Prerequisites, causes and goals of human trust in ai. In Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency (2021), pp. 624–635. [34] Jemmott III, J. B., and Gonzalez, E. Social status, the status distribution, and performance in small groups 1. Journal of Applied Social Psychology 19, 7 (1989), 584–598. [35] Jian, J.-Y., Bisantz, A. M., and Drury, C. G. Foundations for an empirically determined scale of trust in automated systems. International journal of cognitive ergonomics 4, 1 (2000), 53–71. [36] Jiang, J. A., Wade, K., Fiesler, C., and Brubaker, J. R. Supporting serendipity: Opportunities and challenges for human-ai collaboration in qualitative analysis. Proceedings of the ACM on Human-Computer Interaction 5, CSCW1 (2021), 1–23. [37] Kunze, A., Summerskill, S. J., Marshall, R., and Filtness, A. J. Automation transparency: implications of uncertainty communication for human-automation interaction and interfaces. Ergonomics 62, 3 (2019), 345–360. [38] Lambert, F. Family of tesla driver killed in model x crash on autopilot is preparing to sue tesla, Apr 2018. [39] Lee, J., and Moray, N. Trust, control strategies and allocation of function in human-machine systems. Ergonomics 35, 10 (1992), 1243–1270. [40] Lee, J. D., and See, K. A. Trust in automation: Designing for appropriate reliance. Human factors 46, 1 (2004), 50–80. [41] Lewis, J. D., and Weigert, A. Trust as a social reality. Social forces 63, 4 (1985), 967–985. [42] Lewis, M., Sycara, K., and Walker, P. The role of trust in human-robot interaction. In Foundations of trusted autonomy. Springer, Cham, 2018, pp. 135–159. [43] Linton, R. The study of man: an introduction. [44] Lount Jr, R. B., and Pettit, N. C. The social context of trust: The role of status. Organizational Behavior and Human Decision Processes 117, 1 (2012), 15–23. [45] Madhavan, P., and Wiegmann, D. A. Similarities and differences between human-human and human-automation trust: an integrative review. Theoretical Issues in Ergonomics Science 8, 4 (2007), 277–301. [46] Madsen, M., and Gregor, S. Measuring human-computer trust. In 11th australasian conference on information systems (2000), vol. 53, Citeseer, pp. 6–8. [47] Mahmood, A., Fung, J. W., Won, I., and Huang, C.-M. Owning mistakes sincerely: Strategies for mitigating ai errors. [48] Mayer, R. C., Davis, J. H., and Schoorman, F. D. An integrative model of organizational trust. Academy of management review 20, 3 (1995), 709–734. [49] Mayer, R. C., and Gavin, M. B. Trust in management and performance: Who minds the shop while the employees watch the boss? Academy of management journal 48, 5 (2005), 874–888. [50] McAllister, D. J. Affect-and cognition-based trust as foundations for interpersonal cooperation in organizations. Academy of management journal 38, 1 (1995), 24–59. [51] McKnight, D. H., Choudhury, V., and Kacmar, C. Developing and validating trust measures for e-commerce: An integrative typology. Information systems research 13, 3 (2002), 334–359. [52] McKnight, D. H., Cummings, L. L., and Chervany, N. L. Initial trust formation in new organizational relationships. Academy of Management review 23, 3 (1998), 473–490. [53] Niu, Z., Zhou, M., Wang, L., Gao, X., and Hua, G. Ordinal regression with multiple output cnn for age estimation. In Proceedings of the IEEE conference on computer vision and pattern recognition (2016), pp. 4920–4928. [54] Oh, C., Song, J., Choi, J., Kim, S., Lee, S., and Suh, B. I lead, you help but only with enough details: Understanding user experience of co-creation with artificial intelligence. In Proceedings of the 2018 CHI Conference on Human Factors in Computing Systems (2018), pp. 1–13. [55] Okamura, K., and Yamada, S. Adaptive trust calibration for human-ai collaboration. Plos one 15, 2 (2020), e0229132. [56] Parasuraman, R., and Riley, V. Humans and automation: Use, misuse, disuse, abuse. Human factors 39, 2 (1997), 230–253. [57] Ridgeway, C. L., and Berger, J. Expectations, legitimation, and dominance behavior in task groups. American Sociological Review (1986), 603–617. [58] Ross, L. D., Amabile, T. M., and Steinmetz, J. L. Social roles, social control, and biases in social-perception processes. Journal of personality and social psychology 35, 7 (1977), 485. [59] Sande, G. N., Ellard, J. H., and Ross, M. Effect of arbitrarily assigned status labels on self-perceptions and social perceptions: The mere position effect. Journal of Personality and Social Psychology 50, 4 (1986), 684. [60] Schmidt, P., Biessmann, F., and Teubner, T. Transparency and trust in artificial intelligence systems. Journal of Decision Systems 29, 4 (2020), 260–278. [61] Schwaninger, I., Fitzpatrick, G., and Weiss, A. Exploring trust in human-agent collaboration. In Proceedings of 17th European Conference on Computer-Supported Cooperative Work (2019), European Society for Socially Embedded Technologies (EUSSET). [62] Siau, K., and Wang, W. Building trust in artificial intelligence, machine learning, and robotics. Cutter business technology journal 31, 2 (2018), 47–53. [63] Simpson, J. A. Psychological foundations of trust. Current directions in psychological science 16, 5 (2007), 264–268. [64] Skjuve, M., Følstad, A., Fostervold, K. I., and Brandtzaeg, P. B. My chatbot companion-a study of human-chatbot relationships. International Journal of Human-Computer Studies 149 (2021), 102601. [65] Thomas, R., and Skinner, L. Total trust and trust asymmetry: Does trust need to be equally distributed in interfirm relationships? Journal of Relationship Marketing 9, 1 (2010), 43–53. [66] Turner, R. H. Strategy for developing an integrated role theory. Humboldt Journal of Social Relations 7, 1 (1979), 123–139. [67] van den Bosch, K., and Bronkhorst, A. Human-ai cooperation to benefit military decision making. NATO. [68] Vereschak, O., Bailly, G., and Caramiaux, B. How to evaluate trust in ai-assisted decision making? a survey of empirical methodologies. Proceedings of the ACM on Human-Computer Interaction 5, CSCW2 (2021), 1–39. [69] Wang, D., Churchill, E., Maes, P., Fan, X., Shneiderman, B., Shi, Y., and Wang, Q. From human-human collaboration to human-ai collaboration: Designing ai systems that can work together with people. In Extended abstracts of the 2020 CHI conference on human factors in computing systems (2020), pp. 1–6. [70] Xie, T., and Pentina, I. Attachment theory as a framework to understand relationships with social chatbots: A case study of replika. In Proceedings of the 55th Hawaii International Conference on System Sciences (2022). [71] Yu, K., Berkovsky, S., Taib, R., Conway, D., Zhou, J., and Chen, F. User trust dynamics: An investigation driven by differences in system performance. In Proceedings of the 22nd international conference on intelligent user interfaces (2017), pp. 307–317. [72] Zhou,L.,Gao,J.,Li,D.,andShum,H.-Y.Thedesignandimplementationofxiaoice, an empathetic social chatbot. Computational Linguistics 46, 1 (2020), 53–93. |