13,388
Views
14
CrossRef citations to date
0
Altmetric
Research Articles

A Systematic Literature Review of User Trust in AI-Enabled Systems: An HCI Perspective

, , , & ORCID Icon
Pages 1251-1266 | Received 13 Jun 2022, Accepted 14 Oct 2022, Published online: 10 Nov 2022

Abstract

User trust in Artificial Intelligence (AI) enabled systems has been increasingly recognized and proven as a key element to fostering adoption. It has been suggested that AI-enabled systems must go beyond technical-centric approaches and towards embracing a more human-centric approach, a core principle of the human-computer interaction (HCI) field. This review aims to provide an overview of the user trust definitions, influencing factors, and measurement methods from 23 empirical studies to gather insight for future technical and design strategies, research, and initiatives to calibrate the user-AI relationship. The findings confirm that there is more than one way to define trust. Selecting the most appropriate trust definition to depict user trust in a specific context should be the focus instead of comparing definitions. User trust in AI-enabled systems is found to be influenced by three main themes, namely socio-ethical considerations, technical and design features, and user characteristics. User characteristics dominate the findings, reinforcing the importance of user involvement from development through to monitoring of AI-enabled systems. Different contexts and various characteristics of both the users and the systems are also found to influence user trust, highlighting the importance of selecting and tailoring features of the system according to the targeted user group’s characteristics. Importantly, socio-ethical considerations can pave the way in making sure that the environment where user-AI interactions happen is sufficiently conducive to establish and maintain a trusted relationship. In measuring user trust, surveys are found to be the most common method followed by interviews and focus groups. In conclusion, user trust needs to be addressed directly in every context where AI-enabled systems are being used or discussed. In addition, calibrating the user-AI relationship requires finding the optimal balance that works for not only the user but also the system.

1. Introduction

Various real-world applications of Artificial Intelligence (AI) have been developed and implemented to improve, for example, online health platforms (Panda & Mohapatra, Citation2021), banking systems (Mohapatra, Citation2021), businesses, industry (Mohapatra & Kumar, Citation2019), and life in general (Abebe & Goldner, Citation2018; Banerjee et al., Citation2021, Citation2022; Davenport & Ronanki, Citation2018). Nevertheless, with this uptake, more concerns have been raised about certain AI characteristics, such as being opaque (“black box”), uninterpretable, and biased, and the risks these characteristics impose (Gulati et al., Citation2019; Rai, Citation2020). AI opaqueness, for example, makes it harder to predict how AI may behave (Bathaee, Citation2017; Fainman, Citation2019), may make back-tracking errors and decisions more difficult (Bathaee, Citation2017), and the effort to understand the logic of how an output is produced harder (Fainman, Citation2019). These difficulties are amplified with the fact that AI output is suggested to be inherently uncertain (Zhang & Hußmann, Citation2021). When poorly designed or adapted to target users, AI usage could mislead users into unfair and even incorrect decision-making (Lakkaraju & Bastani, Citation2020). Consequently, the real-world consequences of a failed AI-enabled system can be catastrophic, leading to, for example, discrimination (Buolamwini, Citation2017; Buolamwini & Gebru, Citation2018; Dastin, Citation2022; Hoffman & Podgurski, 2022; Kayser-Bril, Citation2020; Olteanu et al., Citation2019; Ruiz, Citation2019), and even death (Kohli & Chadha, Citation2020; Pietsch, Citation2021). Here, AI-enabled systems are defined as AI systems with capabilities to improve existing systems’ performance, i.e., AI-enhanced systems (Boland & Lyytinen, Citation2017), for example, recommender systems, and/or AI systems with capabilities to develop new applications, i.e., AI-based systems (Wuenderlich & Paluch, Citation2017), for example, virtual agents and robotic surgery (Rzepka & Berger, Citation2018).

The potential negative consequences of using AI-enabled systems have led to a lack of trust by users, and highlighted the importance of ethics. As Bryson (Citation2019) stresses, AI-enabled systems are an artifact designed by humans who supposedly should be held accountable for outcomes regardless whether humans understand the logic behind a particular process followed by the systems. As a result, efforts have been put into the creation of AI ethics guidelines to address this issue (Jobin et al., Citation2019). Nevertheless, a review of 84 AI ethics guidelines in different countries reveals that although there are similarities in the principles proposed by these guidelines (i.e., transparency, justice and fairness, non-maleficence, responsibility and privacy) (Jobin et al., Citation2019), there are differences in the interpretation, prioritization, and implementation of the ethical principles. Consequently, AI ethics guidelines can in fact be misplaced and harmful if the efforts distract the focus to operationalize ethical AI-enabled systems (Mittelstadt, Citation2019; Munn, Citation2022). Efforts thus have been broadened to the concept of trustworthy AI that includes AI-enabled systems that are not only ethical, but also lawful and robust (European Commission, Citation2019).

Trustworthiness in AI-enabled systems can be achieved by making sure that the risks associated with certain characteristics of AI-enabled systems are managed (Cheatham et al., Citation2019; Floridi et al., Citation2018). Accordingly, AI developers, researchers and regulators suggest that AI-enabled systems must go beyond technical-centric approaches and towards embracing a more human-centric approach (Shneiderman, Citation2020a), a core principle of the human-computer interaction (HCI) field (Xu, Citation2019; Zhang & Hußmann, Citation2021). According to Hoffman et al. (Citation2001), HCI is a multidisciplinary field of study that focuses on the development, evaluation, and dissemination of technology to meet users’ needs by optimizing how users and technology interact. HCI has broadened its focus during the third wave of computing when technology was embedded into practically all sectors and computers were connected by the internet. Today, HCI covers computer science, engineering, cognitive science, ergonomics, design principles, economics, and behavioral and social sciences to meet rapidly changing user needs (Rogers, Citation2012). Importantly, HCI has been used to ensure trustworthiness of AI in recent years in the form of, for example, guidelines, frameworks, and principles (Leijnen et al., Citation2020; Robert et al., Citation2020; Shneiderman, Citation2020a; Smith, Citation2019). This is because utilizing an HCI approach assumes an interdisciplinary view of technology (Rogers, Citation2012), and thus relies on knowledge from, among others, psychology, sociology, and computer science to develop strategies to foster user trust in AI-enabled systems (Corritore et al., Citation2007; Robert et al., Citation2020).

1.1. Research scope

A growing number of researchers argue that fostering and maintaining user trust is the key to calibrating the user-AI relationship (Jacovi et al., Citation2021; Shin, Citation2021), achieving trustworthy AI (Smith, Citation2019), and further unlocking the potential of AI for society (Bughin et al., Citation2018; Cheatham et al., Citation2019; Floridi et al., Citation2018; Taddeo & Floridi, Citation2018). Despite trust being a concept widely studied across many disciplines (Mcknight & Chervany, Citation1996), the definition, importance and measurement of user trust in AI-enabled systems are not yet as well agreed on and studied (Bauer, Citation2019; Sousa et al., Citation2016). Consequently, the terms trust, trustworthy, and trustworthiness may be addressed without a clear focus and understanding of how these concepts may play a role in the interactions between users and AI-enabled systems, as principles alone cannot guarantee actual trustworthiness (Mittelstadt, Citation2019). Trust is not a direct measure of value and cannot be framed in a single construct. Instead, trust in technology is a social-technical construct and reflects an individual’s willingness to be vulnerable to the actions of another, irrespective of the ability to monitor or control these actions (Mayer et al., Citation2006). In addition, AI-enabled systems are inherently “complex” (Bathaee, Citation2017; Fainman, Citation2019; Mittelstadt, Citation2019), meaning that their functions and design elements make it more challenging for users to immediately understand, accept and justify (Bathaee, Citation2017; Fainman, Citation2019; Mittelstadt, Citation2019). Even when users feel they can control a complex system, they are known to misinterpret the causality of the elements within it (Dörner, Citation1978). These complex characteristics can pose a challenge to address user trust in AI-enabled systems and risk a trust gap between users and the systems (Ashoori & Weisz, Citation2019). The current review focuses on the user-AI relationship because AI is likely to amplify the importance of collaboration between the user and the system, in which the system is one of the collaborative partners in the relationship.

To our knowledge, there is still little research focused on providing an overview of empirical studies focused on user trust in AI-enabled systems where the user-AI relationship is the center point (Xu, Citation2019). Therefore, our study aims to contribute to the HCI literature by providing an overview of user trust definitions, user trust influencing factors, and methods to measure user trust in AI-enabled systems. HCI is naturally used as a search term in this systematic literature review to specifically identify studies that draw on HCI concepts whilst focusing on the user-AI relationship. Findings can be used to provide insight for future technical and design strategies as well as research and initiatives focused on fostering and maintaining user trust in AI-enabled systems.

1.2. Research questions

This systematic literature review aimed to answer: how is user trust in AI-enabled systems defined (RQ1)? What factors influence user trust in AI-enabled systems (RQ2)? How can user trust in AI-enabled systems be measured (RQ3)?

2. Methodology

2.1. Literature search and strategy

Our initial quick search on the topic revealed different approaches to studying user trust in AI-enabled systems. In addition to allowing others to replicate our study, a systematic literature review was chosen because this method provides an overview of the available empirical evidence by reviewing the relevant literature on the chosen topic rigorously and systematically with the aim to minimize bias and produce more reliable results (About Cochrane Reviews, Citationn.d.; Tanveer, Citationn.d.). This method has also been used to, for example, build a theoretical framework for the adoption of AI-enabled systems (Banerjee et al., Citation2021, Citation2022; Panda & Mohapatra, Citation2021).

The systematic literature review was conducted in line with the PRISMA standards for qualitative synthesis (Moher et al., Citation2010). For this purpose, two computer science digital libraries were used: (1) the “EBSCO Discovery Service” and (2) “Web of Science”. The keywords and search strings were selected based on the research questions and a pilot search to ensure all the relevant articles were included. ACM computing classification system (CSS) 2012 was consulted to further refine the search terms strategy (Rous, Citation2012), resulting in the final keywords chosen as shown in .

Table 1. The search terms.

2.2. The study selection

Inclusion and exclusion criteria were developed to define the scope of the study as followed: (1) published in English between January 1, 2011 and May 15, 2021, (2) a scientific article or conference proceeding, (3) empirical, (4) presented factors influencing user trust in AI-enabled systems, (5) presented a clear methodology, and (6) a full-text article version available. The search resulted in a total of 493 articles (). A series of virtual meetings were held to review and discuss the 67 articles included in the full-text screening. Group consensus was used throughout the article selection process to resolve any disagreements regarding eligibility. The first authors (AK, TAB) conducted an additional quality check independently for eligibility. All authors accepted the final 23 articles to be included in the analysis and synthesis ().

Figure 1. PRISMA flow chart of the study selection process.

Figure 1. PRISMA flow chart of the study selection process.

Table 2. Results of the search using EBSCO Discovery Service and Web of Science.

2.3. Analysis and synthesis

Analysis and synthesis were first conducted on three randomly selected articles of the final 23 articles to evaluate the inclusion and exclusion criteria and determine the synthesis process. Authors then independently extracted the following information from the 23 included articles: the title and citation, author(s), the year of publication, the geographical location of the data collection/study, the study focus, the study area/domain/industry, the AI-enabled system being studied, trust definition, methodology, the method to measure user trust, number of participants/dataset, and factors influencing user trust. A series of meetings were held to discuss the process and results of the analyses. Once all authors agreed on the results, the first-authors (AK, TAB) quality checked extracted information from each article for consistency. Common themes of factors influencing user trust were clustered together using content analysis. Group consensus was used throughout the analysis and to resolve any disagreements.

3. Results

Approximately half (52.17%) of the 23 included articles were published after 2018 (). Slightly more than half (56.52%) conducted their studies in the USA and Germany, and 52.17% of the studies were focused on Robotics and E-commerce. The included articles cover various types of AI-enabled systems in which the most common types were general AI/ML and automated algorithms (30.43%). Almost 78.36% articles focused on assessing, predicting, or augmenting antecedents, predictors, critical dimensions and factors of user trust in AI-enabled systems. Moreover, surveys were found to be the most common method to measure user trust (59.56%), either as a standalone or in combination with interviews or focus groups.

Table 3. Overview of the 23 included articles.

3.1. RQ1: How is user trust in AI-enabled systems defined?

Seven articles provided trust definitions (). Eight articles conceptualized trust, but did not define it (Corritore et al., Citation2012; Duffy, Citation2017; Elkins & Derrick, Citation2013; Höddinghaus et al., Citation2021; Law et al., Citation2021; Lee et al., Citation2021; Sharma, Citation2015; Smith, Citation2016), and the remaining eight articles neither defined nor conceptualized trust. Four articles used Mayer’s trust definition (Mayer et al., Citation2006) (Foehr & Germelmann, Citation2020; Glikson & Woolley, Citation2020; Lin et al., Citation2019; Thielsch et al., Citation2018). Two articles (Hoffmann & Söllner, Citation2014; Zhou et al., Citation2020) used Lee and See’s trust definition (Citation2004). One article (Yan et al., Citation2013) developed its own definition in combination with citing trustworthy characteristics from Avizienis et al. (Citation2004).

Table 4. Overview of trust definitions from seven articles.

3.2. RQ2: What factors influence user trust in AI-enabled systems?

Three main themes were identified from the 23 included articles: socio-ethical considerations, technical and design features, and user characteristics (). Eight articles identified socio-ethical considerations, 12 articles identified technical and design features, and 22 articles identified user characteristics.

Table 5. The 23 included articles and factors influencing user trust.

3.2.1. Socio-ethical considerations influencing user trust

Preparing and adjusting the environment where an AI-enabled system was (to be) implemented were suggested as crucial to ensure initial user trust (Lee et al., Citation2021). This was because the development of AI-enabled systems was often faster than the readiness of their potential users, and a mismatch of readiness levels might lead to low user trust. It was suggested to set up mechanisms in place to foster, maintain, and recover user trust (Binmad et al., Citation2017), by, for example, ensuring user data protection (Foehr & Germelmann, Citation2020), encouraging high-quality user interactions (Lin et al., Citation2019), and solution-oriented technical support (Thielsch et al., Citation2018). It was also suggested that user trust was likely to increase over time (Elkins & Derrick, Citation2013). Therefore, building and maintaining open communication with users, for example, by requesting ongoing feedback of an AI-enabled system being used, was suggested as a determinant for user trust (Elkins & Derrick, Citation2013).

Setting up ethical-legal boundaries of AI-enabled systems was consistently seen as a significant challenge due to unclear accountability between involved parties and unclarity of a determinant if harm occurred to the users (O’Sullivan et al., Citation2019). Accountability for any harm to the user was put on the manufacturer if a manufacturing defect was identified, the operator if the use of the system was implicated, or the person responsible for performing the maintenance or adjustment if the system failure was rooted in its maintenance or adjustments (O’Sullivan et al., Citation2019). However, in practice, this pragmatic imputation might not be as simple, for example, in cases where the manufacturer no longer existed, or where the root-cause of damage was unclear (O’Sullivan et al., Citation2019).

3.2.2. Technical and design features influencing user trust

In developing a virtual agent with the purpose to assist and communicate with a user (e.g., chatbots, embodied conversational agents, smart speakers), the following technical and/or design features were found to increase user trust: (1) anthropomorphism and human-like features, especially benevolent features (e.g., smiling, showing interest in the user) in an AI-enabled system (Elkins & Derrick, Citation2013; Foehr & Germelmann, Citation2020; Law et al., Citation2021; Morana et al., Citation2020), (2) immediacy behaviours in which the AI-enabled system could create and project a perception of physical and psychological closeness to the user (Glikson & Woolley, Citation2020), (3) social presence of the AI-enabled system (Glikson & Woolley, Citation2020; Morana et al., Citation2020; Weitz et al., Citation2021), (4) integrity of the AI-enabled system (i.e., repeatedly satisfactory task fulfillment) (Foehr & Germelmann, Citation2020; Höddinghaus et al., Citation2021), (5) additional text/speech output when communicating with users (Weitz et al., Citation2021), (6) providing users with texts rather than a synthetic voice (Law et al., Citation2021), and (7) a lower vocal pitch of the AI-enabled system (Elkins & Derrick, Citation2013).

Specifically for AI/ML and automated algorithms, the following technical and/or design features were found to influence user trust: (1) explanations and information regarding: how the algorithm worked (Glikson & Woolley, Citation2020), AI’s actions (Barda et al., Citation2020; Glikson & Woolley, Citation2020; O’Sullivan et al., Citation2019), reflections of AI reliability (Barda et al., Citation2020), model performance (Zhang, Genc, et al., Citation2021), feature influence methods, risk factors to predictive models (Barda et al., Citation2020), contextual information (Barda et al., Citation2020), and interactive risk explanation tools (baseline risk and risk trends) (Barda et al., Citation2020), (2) correctness of AI/ML predictions (Zhang, Genc, et al., Citation2021), and (3) AI/ML integrity (Höddinghaus et al., Citation2021).

One article, which focused on user trust in complex information systems, found that system reliability (dependability, lack and correctness of data, technical verification, distribution of the system) and the quality of the system information (credibility) influenced user trust (Thielsch et al., Citation2018). Importantly, users with high dependency on the systems and users who had to use the systems, had no other choice but to trust the systems (Thielsch et al., Citation2018). When an information system used a website to interact with users, multimedia features, security certificate/logo, contact information, and a social networking logo were found to be important for user trust (Sharma, Citation2015).

3.2.3. User characteristics influencing user trust

User characteristics dominated the findings and were thus divided by user inherent characteristics (N = 3 articles), user acquired characteristics (N = 4 articles), user attitudes (N = 10 articles), and user external variables (N = 6 articles) ().

3.2.3.1. User inherent characteristics (i.e., personality traits, gender, and self-trust)

Zhou et al. (Citation2020) found that user personality traits influenced user predictive decision making and trust in AI-enabled systems. The study used the big five personality traits (Gosling et al., Citation2003) and found that Low Openness traits (practical, conventional, prefers routine) had the highest trust, followed by Low Conscientiousness (impulsive, careless, disorganized), Low Extraversion (quiet, reserved, withdrawn), and High Neuroticism (anxious, unhappy, prone to negative emotions). Given that personality traits were found to influence user trust, a user interface was suggested to include modules to identify and inform user personality traits to users. This would allow users to be aware of how their personality traits influenced their decision-making when interacting with an AI-enabled system.

Additionally, women were found to be more likely to yield a higher level of trust in an AI-enabled system (Morana et al., Citation2020). Another article looking into user self-trust found that a user was likely to use their own skills to gather and analyze information to decide whether to trust a system (Duffy, Citation2017).

3.2.3.2. User acquired characteristics (i.e., user experiences and educational levels)

A user’s previous experience with a provider or producer of an AI-enabled system was found to influence user trust (Foehr & Germelmann, Citation2020; Yan et al., Citation2013). Positive experiences with a system allowed the user to be rooted deeply in the provider’s or producer’s ecosystem, enabling the transfer of such trust to other systems from the same provider or producer. Importantly, a user’s need or dependency to use a specific AI-enabled system overruled previous negative experiences with the system, especially if the negative experiences were predictable (Yan et al., Citation2013).

Generally, users without a college education were less likely to trust an AI-enabled system than those with a college education (Elkins & Derrick, Citation2013). The study concluded that this finding might be rooted in the unfamiliarity with or the perception of non-benevolence of AI-systems. Nevertheless, the study also found that trust increased over time along with growing familiarity with the system, including when the initial trust level in the AI-enabled system was relatively low.

3.2.3.3. User attitudes (i.e., user acceptance and readiness, needs and expectations, judgment and perceptions)

User acceptance and readiness of an AI-enabled system were found to be key determinants of user trust (Foehr & Germelmann, Citation2020; Khosrowjerdi, Citation2016; Klumpp & Zijm, Citation2019; Smith, Citation2016). Two studies suggested that addressing challenges such as artificial divide (Klumpp & Zijm, Citation2019) and user uncertainties (Hoffmann & Söllner, Citation2014) were fundamental for promoting user acceptance and readiness. The first study defined the artificial divide as the ability or lack thereof to cooperate successfully with AI-enabled systems (Klumpp & Zijm, Citation2019, p. 6). The study outlined that users might be divided by their motivation (e.g., intention to use) and technical competence toward AI-enabled systems. Users with low motivation and low technical competence were the risk group and needed more attention to enable their acceptance and readiness to use AI-enabled systems. The study highlighted the importance of analyzing artificial divide elements (e.g., rejection of an AI-enabled system) and addressing challenges properly (e.g., early stage user involvement, training, enhanced user experience and empowerment) to foster user trust and prevent mistrust (Klumpp & Zijm, Citation2019). The second study suggested that user uncertainties had to be addressed by identifying and prioritizing the uncertainties and their antecedents in relation to a specific AI-enabled system, improving user understandability, sense of control, and information accuracy (Hoffmann & Söllner, Citation2014). A decrease in user uncertainties was suggested as an increase in user trust towards the specific AI-enabled system. Ironically, another study found that perceived imposition or inescapability of AI-enabled systems in general, i.e., a belief that AI-enabled systems would be a part of human daily life nonetheless, would initiate user trust as users perceived that trusting the systems was the only option (Foehr & Germelmann, Citation2020).

User needs and expectations of AI-enabled systems included user intention to use an AI-enabled system (Khosrowjerdi, Citation2016), relevance of technical system quality (e.g., reliability) and information quality (e.g., credibility) to users (Thielsch et al., Citation2018), as well as usefulness of an AI-enabled system to its users (Foehr & Germelmann, Citation2020). In general, user expectations of an AI-enabled system might not be aligned with the intention of the system’s investors and developers (Lee et al., Citation2021). This might result in the system being operated in a way that was unforeseen by investors or developers, hitting and missing the target user expectations. The mismatch between user expectations and experiences was suggested to be a risk to user trust and needed to be addressed, especially when users were heavily dependent on specific AI-enabled systems (Lee et al., Citation2021; Thielsch et al., Citation2018).

For user judgement and perceptions, the key elements found to be affecting user trust in an AI-enabled system included perceived credibility (e.g., expertise, honesty, reputation, and predictability), risk (i.e., likelihood and severity of negative outcomes), and ease of use (e.g., searching, transacting and navigating) (Corritore et al., Citation2012; Foehr & Germelmann, Citation2020) as well as perceived benevolence, integrity and transparency (Elkins & Derrick, Citation2013; Höddinghaus et al., Citation2021). Importantly, it was found that the relatability a user felt to an AI-enabled system determined the user’s trust in the system (Thielsch et al., Citation2018; Zhang, Genc, et al., Citation2021). If user trust was to be fostered, the studies suggested that a focus was needed to increase user relatability to and understandability of an AI-enabled system’s rationale and performance.

3.2.3.4. User external variables (i.e., initial interactions, user interactions, cognitive load levels, time and usage)

When an AI-enabled system was introduced to a potential user through the user’s close relatives, friends or partner, the potential user typically used this opportunity to collect information regarding the system’s benevolence, ability, and integrity (Foehr & Germelmann, Citation2020). In this study, the potential users were aware that they were introduced to an AI-enabled system by their close relatives, friends or partner (Foehr & Germelmann, Citation2020). Importantly, initial trust was likely to be fostered as well. In review-based recommender systems, the quality of user interactions on an AI-enabled system’s platform was found to be a determinant of user trust (Duffy, Citation2017; Lin et al., Citation2019). For example, perceived similarities between users (e.g., preferences and interests) were taken into consideration when evaluating others’ reviews (Duffy, Citation2017; Lin et al., Citation2019). Creating an effective environment where users were willing to exchange social support and share high-quality reviews was suggested as crucial to foster and maintain user trust (Lin et al., Citation2019). Another important determinant of user trust was the user’s cognitive load when interacting with an AI-enabled system (Zhou et al., Citation2020). When under a low cognitive load, the user was more willing to trust a system enabled by a greater availability of the user’s cognitive resources which allowed more confidence and willingness to analyze and understand the AI-enabled system.

One study found that user trust increased as more time was spent interacting with an AI-enabled system (Elkins & Derrick, Citation2013), likely as a result of understanding the system better and thus perceiving it had greater integrity (Elkins & Derrick, Citation2013; Lee et al., Citation2021). The study used an Embodied Conversational Agent (ECA) to ask participants 4 blocks of four questions, a total of 16 questions, similar to those screening questions asked at airports. Participants then were asked to rate their perceived trust of the ECA interviewer after each block of questions. After quantifying the trust ratings, the findings showed that user trust increased after each block of questions, regardless of the initial trust rating. User trust was thus suggested as multidimensional and continuous, and that human-system interactions were crucial to user trust development (Elkins & Derrick, Citation2013). Finally, usage was suggested as a reliable predictor of user trust; the more a user used an AI-enabled system, the more they trusted the system (Yan et al., Citation2013).

3.3. RQ3: How is user trust in AI-enabled systems measured?

A total of 16 studies (69.56%) used a survey either alone or in a combination with an interview or a focus group to measure user trust (). Of the 16 studies, 12 (75%) developed their own questionnaires, two (12.5%) developed their own as well as used previously developed questionnaires, and two (12.5%) used previously developed questionnaires. One of the 12 studies that developed their own questionnaires used it as a pre-survey to collect participant demographic characteristics and one study developed a questionnaire to collect participant preferred design options. Qualitative methods (e.g., interviews and/or focus groups) were the second most common methods used to measure user trust in six studies (26.09%), either as a stand alone or in combination with another method.

Table 6. The 23 included articles’ titles, methods, participants/datasets, and questionnaires.

Twenty articles included participants in their studies, in which the number of participants ranged from 21 to 3423 participants (M = 326.80). Fourteen articles reported gender of the participants (range: 23.08–61.86% females; M = 45.62% females), and two articles reported 4% (Zhang, Genc, et al., Citation2021), 0.51% and 0.71% (Law et al., Citation2021) of gender claimed as other than male or female.

4. Discussion

This systematic literature review has identified 23 empirical studies which investigate how user trust is defined, factors influencing user trust, and methods for measuring user trust in AI-enabled systems. This section will discuss each research question separately.

4.1. RQ1: How is user trust in AI-enabled systems defined?

Of 23 studies, only seven explicitly define trust, while eight conceptualize it and the remaining nine provide neither. This is likely due to trust being an abstract concept that can be relatively difficult to define or generalize (Gebru et al., Citation2022; Gulati et al., Citation2019; Sousa et al., Citation2016), with dynamic characteristics, meaning that trust can change over time and in different contexts and situations (Elkins & Derrick, Citation2013). The difficulty in defining trust is reflected by findings that only one of the 23 included studies develop their own trust definition (Yan et al., Citation2013), whereas six studies use Mayer’s and Lee and See’s trust definitions (Lee & See, Citation2004; Mayer et al., Citation2006).

This finding confirms that there is more than one way to describe trust. Nevertheless, we propose that instead of pursuing better trust definitions or comparing which definitions are better, it is probably more beneficial to select the most appropriate trust definition according to the context, for example, based on the level of risk an output may affect a user. Mayer’s trust definition, for example, may be able to provide a more accurate depiction of user trust in an AI-enabled system in which the output can have a significant personal impact to the user (e.g., personal finance, health). Whereas Lee and See’s may be more accurate to be used for outputs that have a less personal impact to the user (e.g., complex information systems at workplace). The more accurate a trust definition is being used in specific contexts, the easier it is to understand user trust and factors influencing it.

4.2. RQ2: What factors influence user trust in AI-enabled systems

The first key finding is that user characteristics dominate the findings, reinforcing the importance of continuous user involvement from system development through to the implementation and monitoring of AI-enabled systems (Khosrowjerdi, Citation2016; Klumpp & Zijm, Citation2019). The second key finding is that user trust can increase over time due to more user-system interactions (Elkins & Derrick, Citation2013; Lee et al., Citation2021), suggesting that low initial user trust is not fixed and can be improved. This finding highlights the importance of the user-AI interactions as a factor to foster user trust over time (Glikson & Woolley, Citation2020). It is likely that the interactions allow users to adjust their expectations and familiarize themselves to the system, resulting in increased trust in the system (Lee et al., Citation2021). For example, one included study that investigates how stakeholders of a cryptocurrency chatbot experience trust through the bot, found that the participants (users and developers) would start trusting the bot after initial interactions with the bot as part of “a journey” to discover the bot’s features (Lee et al., Citation2021).

The third key finding is that different factors influence user trust based on different contexts and different characteristics of the users and systems. This highlights the importance of selecting and tailoring features of the system according to the targeted user group’s characteristics and attributes. For example, technical and design features found to influence user trust can guide AI-enabled system design strategy (Weitz et al., Citation2021), as well as determining which technical and design features should be emphasized according to contexts and goals of the system tasks (Rheu et al., Citation2021). User characteristics evident to influence user trust can be used to optimize which AI-enabled systems, or their features, fit best for specific types of users. For example, user inherent characteristics can be used as a basis to determine which AI-enabled systems are the best fit and to adjust and improve the system design according to the inherent characteristics of a target group (Duffy, Citation2017; Morana et al., Citation2020; Zhou et al., Citation2020). Whereas, interventions such as user empowerment to foster, maintain or regain user trust in AI-enabled systems can be targeted specifically to improve user attitudes, experiences and external variables (i.e., factors that are more dynamic and open for change than those of user inherent characteristics) (Smith, Citation2019).

The fourth key finding is that socio-ethical considerations can pave the way in making sure that the environment where user-AI interactions happen is sufficiently conducive for these interactions to develop into trusted relationships (O’Sullivan et al., Citation2019). For example, the importance of explanations of AI-enabled systems have been consistently mentioned (Barda et al., Citation2020; Glikson & Woolley, Citation2020; O’Sullivan et al., Citation2019; Zhang, Bengio, et al., Citation2021), highlighting their potential role to improve user experiences and attitudes. Additionally, the findings show that there is still a significant challenge in setting up ethical-legal boundaries for usage of AI-enabled systems due to, for example, a gap between regulations and practices. In this case, a collective, multidisciplinary effort to close this gap is urgently needed (Hagendorff, Citation2020; Shneiderman, Citation2020c).

4.3. RQ3: How user trust in AI-enabled systems is measured

Over two-thirds (69.56%) of the included studies developed and used their own questionnaires to measure user trust (), illustrating that surveys are found to be the most common method to measure user trust. Qualitative methods (e.g., interviews or focus groups) were the second most used methods for measuring trust. Although qualitative methods are the most appropriate method to explore complex topics such as trust (Barda et al., Citation2020; Klumpp & Zijm, Citation2019), the caveat is that results are harder to compare than those of quantitative methods (e.g., surveys) and susceptible to varied interpretations. Nevertheless, these findings highlight different tools to measure user trust and are stated as a concern (Glikson & Woolley, Citation2020; Gulati et al., Citation2019). The concern relates to the fact that if user trust can be understood and measured in different ways, then being able to build upon the concept becomes challenging. A validated tool that allows empirical measurement of user trust across environments and contexts (Schepman & Rodway, Citation2022), such as the 12-item Human-Computer Trust Scale (HCTS) by Gulati et al. (Citation2019), may be used to address this concern. Without doubt, more studies measuring and thus providing a more complete picture of user trust in AI-enabled systems are needed.

4.4. Limitations

This study has several limitations. First, the 23 identified studies have different contexts and some are rather specific. As such, generalization of findings may not be feasible and should be done with caution. Second, it is likely that there are trust definitions other than the ones identified from the 23 included studies. Third, our search terms may be perceived as too narrow and may have resulted in other potentially relevant studies being excluded. Nevertheless, the search terms still capture various types of AI-enabled systems used in different disciplines where the user-AI relationship is the focus. Fourth, our review does not include grey literature, which may have resulted in the exclusion of other potentially relevant work.

5. Conclusions and future aims

The user trust definitions, influencing factors, and measurement methods are crucial topics to be further explored as user trust in AI-enabled systems has been increasingly recognized and proven as a key element to foster adoption (Jacovi et al., Citation2021; Shin, Citation2021). Future studies should investigate and evaluate trust concepts and their applications in specific contexts with AI-enabled systems. Although several factors are found to influence user trust, it is still unclear how these factors fit to distinct types of users and contexts and/or change over time. The factor of time needs to be specifically investigated further and in greater detail to understand how influencing factors play a role at different points of time, and how user-AI interactions evolve into user-AI relationships. Surveys, interviews and focus groups, the most common methods found to measure user trust, are dependent on user perceptions that can be argued as being relatively subjective. Future research may explore other methods, possibly in addition to quantitative or qualitative methods, such as using psychophysiological signals (Barda et al., Citation2020; Gebru et al., Citation2022; Klumpp & Zijm, Citation2019), to gather a more objective insight towards understanding user trust in AI-enabled systems. Approximately half of the studies included in this review were conducted in the USA and Germany, highlighting that future research should be conducted in more diverse geographical locations to help understand how cultural factors influence user trust in AI-enabled systems (Jobin et al., Citation2019; Mohapatra & Kumar, Citation2019; Rheu et al., Citation2021).

Our findings highlight that fostering user trust in AI-enabled systems requires involvement from a multidisciplinary team from the early concept and ideation phases, as has also been suggested by other similar studies (Mohapatra, Citation2021; Panda & Mohapatra, Citation2021). Such a team should involve not only AI-enabled system developers, designers and target users, but also individuals with ethics, legal, behavioral, social sciences, and domain/industry expertise (Dwivedi et al., Citation2021). Integrating target user characteristics into technical and design development of AI-enabled systems needs to consider which characteristics are inherent, acquired, attitudes and external variables. This is because inherent user characteristics, for example, are less likely to change compared to user attitudes, thus requiring the systems to be adjusted accordingly. In contrast, user attitudes and external variables, for example, are probably more responsive to efforts to improve user experience (Zhang, Genc, et al., Citation2021). Ensuring quality interactions between users and AI-enabled systems requires adjusting the environment where these interactions happen by, for example, setting up mechanisms to foster, maintain and recover user trust. Importantly, efforts to calibrate the user-AI relationship requires finding the optimal balance that works for not only the user but also the system (DiSalvo et al., Citation2002; Fink, Citation2012; Gebru et al., Citation2022; Shneiderman, Citation2020b). This is because “we still believe that robots – as well as humans – need to be authentic in the way they are, to be ‘successful’ in a variety of dimensions.” (Fink, Citation2012, p. 205).

Author contributions

SS developed the concept and idea. All authors performed data collection, analysis, and interpretation. TAB and AK quality checked the results and interpretation for consistency and led the writing of the manuscript. All authors contributed to the writing and agreed with the final version of the manuscript.

Disclosure statement

No potential conflict of interest was reported by the author(s).

Additional information

Funding

This work was supported by AI-Mind that has received funding from the European Union’s Horizon 2020 research and innovation programme under grant agreement No 964220. This study was partly funded by the Trust and Influence Programme [FA8655-22-1-7051], European Office of Aerospace Research and Development, and US Air Force Office of Scientific Research.

Notes on contributors

Tita Alissa Bach

Tita Alissa Bach holds a doctorate degree in behavioural and social sciences from Groningen University, the Netherlands, with a strong focus in Human Factors and Organisational Psychology. Her research focuses on applying Human Factors principles in an environment or organization, especially in the context where technology is being used.

Amna Khan

Amna Khan is a Junior Researcher and PhD student in Information Society Technologies at Tallinn University in Estonia. Her research focuses on the creation of a framework for an adaptive system to enhance knowledge reuse to develop trustworthy and transparent systems. She earned a master’s degree in Mechatronics Engineering (2017).

Harry Hallock

Harry Hallock holds his doctorate degree in Cognitive Neurosciences from the University of Sydney, Australia. He has experience in product development and large-scale project management. His current work focuses on understanding trustworthy AI in healthcare, federated networks for health data, governance, management and quality assurance of health data.

Gabriela Beltrão

Gabriela Beltrão candidate in Information Society Technologies at Tallinn University, Estonia. Her research targets trust in technology from a human-centered perspective, focusing on differences in trust across cultures and their implications for the design.

Sonia Sousa

Sonia Sousa is an Associate Professor of Interaction Design, Tallinn University, Estonia, in HCI, and the head of the Join Online MSc in Interaction Design. Her funded projects include NGI-Trust, CHIST-ERA, Horizon 2020, and AFOSR. She was nominated for AcademiaNet - the expert database for outstanding female academics.

References

  • Abebe, R., & Goldner, K. (2018). Mechanism design for social good. AI Matters, 4(3), 27–34. https://doi.org/10.1145/3284751.3284761
  • About Cochrane Reviews. ( n.d.). Cochranelibrary. https://www.cochranelibrary.com/about/about-cochrane-reviews
  • Ashoori, M., Weisz, J. D. (2019). In AI we trust? Factors that influence trustworthiness of AI-infused decision-making processes. In arXiv [cs.CY]. arXiv. http://arxiv.org/abs/1912.02675
  • Avizienis, A., Laprie, J.-C., Randell, B., & Landwehr, C. (2004). Basic concepts and taxonomy of dependable and secure computing. IEEE Transactions on Dependable and Secure Computing, 1(1), 11–33. https://doi.org/10.1109/tdsc.2004.2
  • Banerjee, S. S., Mohapatra, S., & Saha, G. (2021) Developing a framework of artificial intelligence for fashion forecasting and validating with a case study. International Journal of High Risk Behaviors & Addiction, 12(2), 165–180. https://www.researchgate.net/profile/Sanjay-Mohapatra/publication/353277720_Developing_a_framework_of_artificial_intelligence_for_fashion_forecasting_and_validating_with_a_case_study/links/61e80a065779d35951bcbb96/Developing-a-framework-of-artificial-intelligence-for-fashion-forecasting-and-validating-with-a-case-study.pdf
  • Banerjee, S., Mohapatra, S., & Bharati, M. (2022). AI in fashion industry. Emerald Group Publishing. https://doi.org/10.1108/9781802626339
  • Barda, A. J., Horvat, C. M., & Hochheiser, H. (2020). A qualitative research framework for the design of user-centered displays of explanations for machine learning model predictions in healthcare. BMC Medical Informatics and Decision Making, 20(1), 257. https://doi.org/10.1186/s12911-020-01276-x
  • Bathaee, Y. (2017). The artificial intelligence black box and the failure of intent and causation. Harvard Journal of Law & Technology, 31, 889.
  • Bauer, P. C. (2019). Clearing the jungle: Conceptualizing trust and trustworthiness. https://doi.org/10.2139/ssrn.2325989
  • Binmad, R., Li, M., Wang, Z., Deonauth, N., & Carie, C. A. (2017). An extended framework for recovering from trust breakdowns in online community settings. Future Internet, 9(3), 36. https://doi.org/10.3390/fi9030036
  • Boland, R. J., & Lyytinen, K. (2017). The limits to language in doing systems design. European Journal of Information Systems, 26(3), 248–259. https://doi.org/10.1057/s41303-017-0043-4
  • Bryson, J. (2019.). The artificial intelligence of the ethics of artificial intelligence. In M. D. Dubber, F. Pasquale & S. Das (Eds), The Oxford handbook of ethics of AI. Oxford University Press. https://books.google.com/books?hl=en&lr=&id=8PQTEAAAQBAJ&oi=fnd&pg=PA3&dq=he+Artificial+Intelligence+of+the+Ethics+of+Artificial+Intelligence+An+Introductory+Overview+for+Law+and+Regulation+The+Oxford+Handbook+of+Ethics+of+Artificial+Intelligence&ots=uCdAtk0bYC&sig=AdrFxuDqwzebzIMSmMhP_OOlI6o
  • Bughin, J., Seong, J., Manyika, J., Chui, M., Joshi, R. (2018). Notes from the AI frontier: Modeling the impact of AI on the world economy. McKinsey Global Institute. https://www.mckinsey.com/∼/media/McKinsey/Featured%20Insights/Artificial%20Intelligence/Notes%20from%20the%20frontier%20Modeling%20the%20impact%20of%20AI%20on%20the%20world%20economy/MGI-Notes-from-the-AI-frontier-Modeling-the-impact-of-AI-on-the-world-economy-September-2018.pdf
  • Buolamwini, J. A. (2017). Gender shades: Intersectional phenotypic and demographic evaluation of face datasets and gender classifiers. Massachusetts Institute of Technology.
  • Buolamwini, J., & Gebru, T. (2018). Gender shades: Intersectional accuracy disparities in commercial gender classification. In S. A. Friedler & C. Wilson (Eds.), Proceedings of the 1st conference on fairness, accountability and transparency (Vol. 81, pp. 77–91). PMLR.
  • Caruso, D. R., & Salovey, P. (2004). The emotionally intelligent manager: How to develop and use the four key emotional skills of leadership. John Wiley & Sons.
  • Cheatham, B., Javanmardian, K., & Samandari, H. (2019). Confronting the risks of artificial intelligence. McKinsey Quarterly, 2, 38. https://www.mckinsey.com/capabilities/quantumblack/our-insights/confronting-the-risks-of-artificial-intelligence
  • Corritore, C. L., Wiedenbeck, S., Kracher, B., & Marble, R. (2007). Online trust and health information websites. In Proceedings of the 6th Annual Workshop on HCI Research in MIS (pp. 25–29). IGI Global United States.
  • Corritore, C. L., Wiedenbeck, S., Kracher, B., & Marble, R. P. (2012). Online trust and health information websites. International Journal of Technology and Human Interaction, 8(4), 92–115. https://doi.org/10.4018/jthi.2012100106
  • Dastin, J. (2022). Amazon scraps secret AI recruiting tool that showed bias against women. In Ethics of data and analytics (pp. 296–299). Auerbach Publications.
  • Davenport, T. H., & Ronanki, R. (2018). Artificial intelligence for the real world. Harvard Business Review, 96(1), 108–116.https://hbr.org/webinar/2018/02/artificial-intelligence-for-the-real-world
  • DiSalvo, C. F., Gemperle, F., Forlizzi, J., & Kiesler, S. (2002). All robots are not created equal: The design and perception of humanoid robot heads. In Proceedings of the 4th Conference on Designing Interactive Systems: Processes, Practices, Methods, and Techniques, (pp. 321–326). ACM.
  • Dörner, D. (1978). Theoretical advances of cognitive psychology relevant to instruction. In A. M. Lesgold, J. W. Pellegrino, S. D. Fokkema, & R. Glaser (Eds.), Cognitive psychology and instruction (pp. 231–252). Springer US.
  • Duffy, A. (2017). Trusting me, trusting you: Evaluating three forms of trust on an information-rich consumer review website. Journal of Consumer Behaviour, 16(3), 212–220. https://doi.org/10.1002/cb.1628
  • Dwivedi, Y. K., Hughes, L., Ismagilova, E., Aarts, G., Coombs, C., Crick, T., Duan, Y., Dwivedi, R., Edwards, J., Eirug, A., Galanos, V., Ilavarasan, P. V., Janssen, M., Jones, P., Kar, A. K., Kizgin, H., Kronemann, B., Lal, B., Lucini, B., … Williams, M. D. (2021). Artificial Intelligence (AI): Multidisciplinary perspectives on emerging challenges, opportunities, and agenda for research, practice and policy. International Journal of Information Management, 57, 101994. https://doi.org/10.1016/j.ijinfomgt.2019.08.002
  • Ehsan, U., Tambwekar, P., Chan, L., Harrison, B., & Riedl, M. O. (2019). Automated rationale generation: A technique for explainable AI and its effects on human perceptions. In Proceedings of the 24th International Conference on Intelligent User Interfaces (pp. 263–274). ACM.
  • Elkins, A. C., & Derrick, D. C. (2013). The sound of trust: Voice as a measurement of trust during interactions with embodied conversational agents. Group Decision and Negotiation, 22(5), 897–913. https://doi.org/10.1007/s10726-012-9339-x
  • Fainman, A. A. (2019). The problem with opaque AI. The Thinker, 82(4), 44–55. https://doi.org/10.36615/thethinker.v82i4.373
  • Fink, J. (2012). Anthropomorphism and human likeness in the design of robots and human-robot interaction. In Social robotics (pp. 199–208). Springer. https://doi.org/10.1007/978-3-642-34103-8_20
  • Floridi, L., Cowls, J., Beltrametti, M., Chatila, R., Chazerand, P., Dignum, V., Luetge, C., Madelin, R., Pagallo, U., Rossi, F., Schafer, B., Valcke, P., & Vayena, E. (2018). AI4People—an ethical framework for a good AI society: Opportunities, risks, principles, and recommendations. Minds and Machines, 28(4), 689–707. https://doi.org/10.1007/s11023-018-9482-5
  • Foehr, J., & Germelmann, C. C. (2020). Alexa, can i trust you? Exploring consumer paths to trust in smart voice-interaction technologies. Journal of the Association for Consumer Research, 5(2), 181–205. https://doi.org/10.1086/707731
  • Gebru, B., Zeleke, L., Blankson, D., Nabil, M., Nateghi, S., Homaifar, A., & Tunstel, E. (2022). A review on human–machine trust evaluation: Human-centric and machine-centric perspectives. IEEE Transactions on Human-Machine Systems, 52(5), 1–11. https://doi.org/10.1109/THMS.2022.3144956
  • Gefen, D., Benbasat, I., & Pavlou, P. (2008). A research agenda for trust in online environments. In Journal of Management Information Systems, 24(4), 275–286. https://doi.org/10.2753/mis0742-1222240411
  • Glikson, E., & Woolley, A. W. (2020). Human trust in artificial intelligence: Review of empirical research. Academy of Management Annals, 14(2), 627–660. https://doi.org/10.5465/annals.2018.0057
  • Gosling, S. D., Rentfrow, P. J., & Swann, W. B. (2003). A very brief measure of the Big-Five personality domains. Journal of Research in Personality, 37(6), 504–528. https://doi.org/10.1016/S0092-6566(03)00046-1
  • Gulati, S., Sousa, S., & Lamas, D. (2019). Design, development and evaluation of a human-computer trust scale. Behaviour & Information Technology, 38(10), 1004–1015. https://doi.org/10.1080/0144929x.2019.1656779
  • Hagendorff, T. (2020). The ethics of AI ethics: An evaluation of guidelines. Minds and Machines, 30(1), 99–120. https://doi.org/10.1007/s11023-020-09517-8
  • European Commission. (2019). High-level expert group on artificial intelligence. European Commission. https://ec.europa.eu/digital-single, https://42.cx/wp-content/uploads/2020/04/AI-Definition-EU.pdf
  • Höddinghaus, M., Sondern, D., & Hertel, G. (2021). The automation of leadership functions: Would people trust decision algorithms? Computers in Human Behavior, 116, 106635. https://doi.org/10.1016/j.chb.2020.106635
  • Hoffmann, H., & Söllner, M. (2014). Incorporating behavioral trust theory into system development for ubiquitous applications. Personal and Ubiquitous Computing, 18(1), 117–128. https://doi.org/10.1007/s00779-012-0631-1
  • Hoffman, S., & Podgurski, A. (2020). Artificial intelligence and discrimination in health care. Yale Journal of Health Policy, Law, and Ethics. 18(1). https://doi.org/10.1007/s00779-012-0631-1
  • Hoffman, R. R., Hayes, P. J., & Ford, K. M. (2001). Human-centered computing: Thinking in and out of the box. IEEE Intelligent Systems, 16(5), 76–78. https://doi.org/10.1109/MIS.2001.956085
  • Jacovi, A., Marasović, A., Miller, T., & Goldberg, Y. (2021). Formalizing trust in artificial intelligence: Prerequisites, causes and goals of human trust in AI. In Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency (pp. 624–635). ACM.
  • Jian, J.-Y., Bisantz, A. M., & Drury, C. G. (2000). Foundations for an empirically determined scale of trust in automated systems. International Journal of Cognitive Ergonomics, 4(1), 53–71. https://doi.org/10.1207/S15327566IJCE0401_04
  • Jobin, A., Ienca, M., & Vayena, E. (2019). Artificial intelligence: The global landscape of ethics guidelines. Nature Machine Intelligence, 1, 389–399. https://doi.org/10.1038/s42256-019-0088-2
  • Kayser-Bril, N. (2020). Google apologizes after its Vision AI produced racist results. AlgorithmWatch. Retrieved August, 17, 2020. https://algorithmwatch.org/en/google-vision-racism/
  • Khosrowjerdi, M. (2016). A review of theory-driven models of trust in the online health context. IFLA Journal, 42(3), 189–206. https://doi.org/10.1177/0340035216659299
  • Klumpp, M., & Zijm, H. (2019). Logistics innovation and social sustainability: How to prevent an artificial divide in human–computer interaction. Journal of Business Logistics, 40(3), 265–278. https://doi.org/10.1111/jbl.12198
  • Kohli, P., & Chadha, A. (2020). Enabling pedestrian safety using computer vision techniques: A Case study of the 2018 Uber Inc. Self-driving car crash. In Lecture notes in networks and systems (pp. 261–279). Springer. https://doi.org/10.1007/978-3-030-12388-8_19
  • Lakkaraju, H., & Bastani, O. (2020). “How do I fool you?”: Manipulating user trust via misleading black box explanations. In Proceedings of the AAAI/ACM Conference on AI, Ethics, and Society (pp. 79–85). Association for Computing Machinery.
  • Law, T., Chita-Tegmark, M., & Scheutz, M. (2021). The interplay between emotional intelligence, trust, and gender in human–robot interaction. International Journal of Social Robotics, 13(2), 297–309. https://doi.org/10.1007/s12369-020-00624-1
  • Lee, J. D., & See, K. A. (2004). Trust in automation: Designing for appropriate reliance. Human Factors, 46(1), 50–80. https://doi.org/10.1518/hfes.46.1.50_30392
  • Lee, M., Frank, L., & IJsselsteijn, W. (2021). Brokerbot: A cryptocurrency chatbot in the social-technical gap of trust. Computer Supported Cooperative Work (CSCW), 30(1), 79–117. https://doi.org/10.1007/s10606-021-09392-6
  • Leijnen, S., Aldewereld, H., van Belkom, R., Bijvank, R., & Ossewaarde, R. (2020). An agile framework for trustworthy AI. NeHuAI@ ECAI, 75–78.
  • Lin, X., Wang, X., & Hajli, N. (2019). Building E-commerce satisfaction and boosting sales: The role of social commerce trust and its antecedents. International Journal of Electronic Commerce, 23(3), 328–363. https://doi.org/10.1080/10864415.2019.1619907
  • Mayer, R. C., & Davis, J. H. (1999). The effect of the performance appraisal system on trust for management: A field quasi-experiment. Journal of Applied Psychology, 84(1), 123–136. https://doi.org/10.1037/0021-9010.84.1.123
  • Mayer, R. C., Davis, J. H., & Schoorman, F. D. (2006). An integrative model of organizational trust. Organizational Trust: A Reader, 20(2), 82–108. https://doi.org/10.5465/amr.1995.9508080335
  • Mcknight, D., & Chervany, N. (1996). The meanings of trust. https://citeseerx.ist.psu.edu/viewdoc/summary?doi=10.1.1.155.1213
  • Meeßen, S. M., Thielsch, M. T., & Hertel, G. (2020). Trust in management information systems (MIS). Zeitschrift für Arbeits- und Organisationspsychologie A&O, 64(1), 6–16. https://doi.org/10.1026/0932-4089/a000306
  • Mittelstadt, B. (2019). Principles alone cannot guarantee ethical AI. Nature machine intelligence, 1(11), 501–507. https://doi.org/10.1038/s42256-019-0114-4
  • Mohapatra, S. (2021). Human and computer interaction in information system design for managing business. Information Systems and e-Business Management, 19(1), 1–11. https://doi.org/10.1007/s10257-020-00475-3
  • Mohapatra, S., & Kumar, A. (2019). Developing a framework for adopting artificial intelligence. International Journal of Computer Theory and Engineering, 11(2), 19–22. https://doi.org/10.7763/IJCTE.2019.V11.1234
  • Moher, D., Liberati, A., Tetzlaff, J., & Altman, D. G. (2010). Preferred reporting items for systematic reviews and meta-analyses: The PRISMA statement. International Journal of Surgery (London, England), 8(5), 336–341.
  • Morana, S., Gnewuch, U., Jung, D., & Granig, C. (2020). The effect of anthropomorphism on investment decision-making with robo-advisor chatbots. ECIS. https://www.researchgate.net/profile/Stefan-Morana/publication/341277570_The_Effect_of_Anthropomorphism_on_Investment_Decision-Making_with_Robo-Advisor_Chatbots/links/5eb7c5ba4585152169c14505/The-Effect-of-Anthropomorphism-on-Investment-Decision-Making-with-Robo-Advisor-Chatbots.pdf
  • Munn, L. (2022). The uselessness of AI ethics. AI and Ethic. 1–9. https://doi.org/10.1007/s43681-022-00209-w
  • Olteanu, A., Castillo, C., Diaz, F., & Kıcıman, E. (2019). Social data: Biases, methodological pitfalls, and ethical boundaries. Frontiers in Big Data, 2, 13.
  • O’Sullivan, S., Nevejans, N., Allen, C., Blyth, A., Leonard, S., Pagallo, U., Holzinger, K., Holzinger, A., Sajid, M. I., & Ashrafian, H. (2019). Legal, regulatory, and ethical frameworks for development of standards in artificial intelligence (AI) and autonomous robotic surgery. The International Journal of Medical Robotics + Computer Assisted Surgery : MRCAS, 15(1), e1968.
  • Panda, A., & Mohapatra, S. (2021). Online healthcare practices and associated stakeholders: Review of literature for future research agenda. Vikalpa: The Journal for Decision Makers, 46(2), 71–85. https://doi.org/10.1177/02560909211025361
  • Pietsch, B. (2021). Killed in driverless Tesla car crash, officials Say. The New York Times.
  • Rai, A. (2020). Explainable AI: From black box to glass box. Journal of the Academy of Marketing Science, 48(1), 137–141. https://doi.org/10.1007/s11747-019-00710-5
  • Rheu, M., Shin, J. Y., Peng, W., & Huh-Yoo, J. (2021). Systematic review: Trust-building factors and implications for conversational agent design. International Journal of Human–Computer Interaction, 37(1), 81–96. https://doi.org/10.1080/10447318.2020.1807710
  • Robert, L. P., Bansal, G., & Lütge, C. (2020). ICIS 2019 SIGHCI workshop panel report: Human– computer interaction challenges and opportunities for fair, trustworthy and ethical artificial intelligence. AIS Transactions on Human-Computer Interaction, 12(2), 96–108. https://doi.org/10.17705/1thci.00130
  • Rogers, Y. (2012). HCI theory: Classical, modern, and contemporary. Synthesis Lectures on Human-Centered Informatics, 5(2), 1–129. https://doi.org/10.2200/s00418ed1v01y201205hci014
  • Rous, B. (2012). Major update to ACM’s computing classification system. Communications of the ACM, 55(11), 12–12. https://doi.org/10.1145/2366316.2366320
  • Ruiz, C. (2019). Leading online database to remove 600,000 images after art project reveals its racist bias. The Art Newspaper, 23. https://authenticationinart.org/wp-content/uploads/2019/09/Leading-online-database-to-remove-600000-images-after-art-project-reveals-its-racist-bias-The-Art-Newspaper.pdf
  • Rzepka, C., & Berger, B. (2018). User interaction with AI-enabled systems: A systematic review of IS research. In Thirty Ninth International Conference on Information Systems (pp. 1–17). Association for Information Systems (AIS).
  • Schepman, A., & Rodway, P. (2022). The general attitudes towards artificial intelligence scale (GAAIS): Confirmatory validation and associations with personality, corporate distrust, and general trust. International Journal of Human–Computer Interaction, 1–18. https://doi.org/10.1080/10447318.2022.2085400
  • Schlosser, A. E., White, T. B., & Lloyd, S. M. (2006). Converting Web site visitors into buyers: How Web site investment increases consumer trusting beliefs and online purchase intentions. Journal of Marketing, 70(2), 133–148. https://doi.org/10.1509/jmkg.70.2.133
  • Sharma, V. M. (2015). A comparison of consumer perception of trust-triggering appearance features on Indian group buying websites. Indian Journal of Economics and Business; New Delhi 14(2), 163–177. https://search.proquest.com/openview/17dbb5773788eb3bd3cd998ea624618a/1?pq-origsite=gscholar&cbl=2026690
  • Shin, D. (2021). The effects of explainability and causability on perception, trust, and acceptance: Implications for explainable AI. International Journal of Human-Computer Studies, 146, 102551. https://doi.org/10.1016/j.ijhcs.2020.102551
  • Shneiderman, B. (2020a). Human-centered artificial intelligence: Three fresh ideas. AIS Transactions on Human-Computer Interaction, 12(3), 109–124. https://doi.org/10.17705/1thci.00131
  • Shneiderman, B. (2020b). Human-centered artificial intelligence: Reliable, safe & trustworthy. International Journal of Human–Computer Interaction, 36(6), 495–504. https://doi.org/10.1080/10447318.2020.1741118
  • Shneiderman, B. (2020c). Bridging the gap between ethics and practice: Guidelines for reliable, safe, and trustworthy human-centered AI systems. ACM Transactions on Interactive Intelligent Systems, 10(4), 1–31. https://doi.org/10.1145/3419764
  • Smith, C. J. (2019). Designing Trustworthy AI: A Human-Machine Teaming Framework to Guide Development. In arXiv [cs.AI]. arXiv. http://arxiv.org/abs/1910.03515
  • Smith, C. L. (2016). Privacy and trust attitudes in the intent to volunteer for data-tracking research. Information Research: An International Electronic Journal, 21(4), 4. https://doi.org/10.1186/s40660-015-0004-y
  • Sousa, S., Lamas, D., & Dias, P. (2016). Value creation through trust in technological-mediated social participation. Technology, Innovation and Education, 2(1), 1–9. https://doi.org/10.1186/s40660-016-0011-7
  • Taddeo, M., & Floridi, L. (2018). How AI can be a force for good. Science, 361(6404), 751–752. https://doi.org/10.1126/science.aat5991
  • Tanveer, T. P. P. (n.d.). Why systematic reviews matter. Retrieved February 9, 2022, from https://www.elsevier.com/connect/authors-update/why-systematic-reviews-matter
  • Thielsch, M. T., Meeßen, S. M., & Hertel, G. (2018). Trust and distrust in information systems at the workplace. PeerJ, 6, e5483. https://doi.org/10.7717/peerj.5483
  • Ullman, D., & Malle, B. F. (2018). What does it mean to trust a robot?. In Steps toward a multidimensional measure of trust. Companion of the 2018 ACM/IEEE International Conference on Human-Robot Interaction (pp. 263–264). ACM.
  • Wang, Y. D., & Emurian, H. H. (2005). An overview of online trust: Concepts, elements, and implications. Computers in Human Behavior, 21(1), 105–125. https://doi.org/10.1016/j.chb.2003.11.008
  • Weitz, K., Schiller, D., Schlagowski, R., Huber, T., & André, E. (2021). “Let me explain!”: Exploring the potential of virtual agents in explainable AI interaction design. Journal on Multimodal User Interfaces, 15(2), 87–98. https://doi.org/10.1007/s12193-020-00332-0
  • Wuenderlich, N. V., & Paluch, S. (2017). A nice and friendly chat with a bot: User perceptions of AI-based service agents. In ICIS 2017 Proceedings. Association for Information Systems (AIS). https://aisel.aisnet.org/icis2017/ServiceScience/Presentations/11/
  • Xu, W. (2019). Toward human-centered AI: A perspective from human-computer interaction. Interactions, 26(4), 42–46. https://doi.org/10.1145/3328485
  • Yan, Z., Dong, Y., Niemi, V., & Yu, G. (2013). Exploring trust of mobile applications based on user behaviors: An empirical study. Journal of Applied Social Psychology, 43(3), 638–659. https://doi.org/10.1111/j.1559-1816.2013.01044.x
  • Yan, Z., Niemi, V., Dong, Y., & Yu, G. (2008). A user behavior based trust model for mobile applications. Autonomic and Trusted Computing, 455–469.
  • Zhang, C., Bengio, S., Hardt, M., Recht, B., & Vinyals, O. (2021). Understanding deep learning (still) requires rethinking generalization. Communications of the ACM, 64(3), 107–115. https://doi.org/10.1145/3446776
  • Zhang, Z., Genc, Y., Wang, D., Ahsen, M. E., & Fan, X. (2021). Effect of AI explanations on human perceptions of patient-facing AI-powered healthcare systems. Journal of Medical Systems, 45(6), 64.
  • Zhang, Z. T., & Hußmann, H. (2021). How to manage output uncertainty: Targeting the actual end user problem in interactions with AI. In Joint Proceedings of the ACM IUI 2021 Workshops Co-Located with the 26th ACM Conference on Intelligent User Interfaces, to Appear, College Station, TX, (IUI’21). Aachen: R. Piskac c/o Redaktion Sun SITE, Informatik V, RWTH Aachen. http://ceur-ws.org/Vol-2903/IUI21WS-TExSS-17.pdf
  • Zhou, J., Luo, S., & Chen, F. (2020). Effects of personality traits on user trust in human–machine collaborations. Journal on Multimodal User Interfaces, 14(4), 387–400. https://doi.org/10.1007/s12193-020-00329-9