6,210
Views
1
CrossRef citations to date
0
Altmetric
Research Article

Toward a design theory for virtual companionship

ORCID Icon, ORCID Icon, ORCID Icon & ORCID Icon
Pages 194-234 | Received 08 Oct 2021, Accepted 23 May 2022, Published online: 18 Jul 2022

1. Introduction and motivation

Due to significant technological advances in the field of artificial intelligence (AI), which are driven by increased computing power, the ubiquitous availability of data, as well as new algorithms, new forms of intelligent systems and services have been developed and brought to the market (Choudhury et al., Citation2020; Clark et al., Citation2019a; Diederich et al., Citation2022; Kaplan & Haenlein, Citation2019; Ransbotham et al., Citation2018; Robert et al., Citation2020; Rzepka & Berger, Citation2018). In addition to specific applications in the form of virtual assistants, such as Apple’s Siri or Amazon’s Alexa, companies increasingly develop chatbots and enterprise bots to interact with customers (Diederich et al., Citation2022; Maedche et al., Citation2016; McTear et al., Citation2016). What all these systems have in common, is that they allow their users to interact with them using natural language, which is why the systems are summarized by the term conversational agent (CA) (Diederich et al., Citation2022; McTear et al., Citation2016). There are already various use cases for CAs today, ranging from executing smartphone functions, such as creating calendar entries or sending messages to smart home control, to interaction in the healthcare context (Ahmad et al., Citation2022; Elshan et al., Citation2022; Gnewuch et al., Citation2017; McTear et al., Citation2016; Sin & Munteanu, Citation2020). Thus, CAs currently offer a new way of interacting with information technology (Morana et al., Citation2017). Recent literature reviews show a growing interest in CAs and AI-enabled systems (Diederich et al., Citation2022; Elshan et al., Citation2022; Nißen et al., Citation2021; Rzepka & Berger, Citation2018), but mainly a limited variety of application contexts, which mostly focus on short-term interactions in marketing, sales, and support. Application scenarios that require long-term interaction are available but under-researched (Diederich et al., Citation2022; Elshan et al., Citation2022). Additionally, the current applications show that CA’s main goal is to provide personal assistant functionality, while less attention goes to the actual interaction with the system which should be improved by social behaviors being incorporated (Elshan et al., Citation2022; Gnewuch et al., Citation2017; Krämer et al., Citation2011; Nißen et al., Citation2021; Rzepka & Berger, Citation2018). Most of these interactions are initiated by the user and not by the CA, which means that the CA acts reactively rather than proactively. Moreover, these interactions are isolated, transactional, and based on predefined paths, as if they are starting over every time (Seymour et al., Citation2018). Although presently, from a technological perspective, CAs can predominantly conduct restricted conversations related to a specific topic (Diederich et al., Citation2022), modern language prediction models such as the Generative Pre-trained Transformer 3 (GPT-3) are able to fundamentally expand the capabilities of CAs. They achieve this by enabling open-topic and richer conversations with strong interpersonal character (Brown et al., Citation2020). The GPT-3 and many other recent language models are built on Transformer (Vaswani et al., Citation2017), a neural network architecture invented by Google Research in 2017. Google’s recent language model LaMDA shows how human-like ways of interacting can be achieved by enabling open-topic conversations based on the modern autoregressive language model (a Transformer-based deep learning neural network). Google argues that although conversations “tend to revolve around a specific topic, their open-ended nature means they can start in one place and end up somewhere completely different.”Footnote1 Especially when a person is talking to a friend instead of an assistant, the open-ended nature of conversation is highly important to ensure a co-equal discussion (Collins & Ghahramani, Citation2021).

As research has shown, humans often treat machines like social actors and mindlessly apply social rules and expectations to them (Nass & Moon, Citation2000; Nass et al., Citation1994). If machines are designed to act in a human-like manner, interpersonal relationships can be established when long-term interactions with them take place. Thus, incorporating theories of interpersonal relationships in the design of human-machine interaction can lead to more natural interaction and prompt requirements for a collaborative and benevolent relationship between humans and machines (Krämer et al., Citation2011; Lee et al., Citation2021; Qiu & Benbasat, Citation2009). Prior research has already discussed different characteristics for establishing long-term relationships between humans and machines by incorporating findings from social interaction studies (Bickmore & Picard, Citation2005; Danilava et al., Citation2012; Krämer et al., Citation2011; Seymour et al., Citation2018). Research shows that social relationships are important conditions that contribute to psychological well-being (Krämer et al., Citation2011). Companionship, a special form of friendship, is perceived as an important aspect to strengthen psychological well-being (Rook, Citation1987). Companionship is described as having someone familiar with whom you like spending time, while a relationship is characterized by an inner and intimate bond, an intense affection, and a high emotional connection (Bukowski et al., Citation1993, Citation1994; Mendelson & Aboud, Citation2012). Therefore, companions differ significantly from assistants, who are not co-equal, do not elicit feelings of closeness, do not form long-term social bonds, and do not behave proactively (Bukowski et al., Citation1994; Krämer et al., Citation2011; Seymour et al., Citation2018). In human-computer interaction, many researchers believe that systems with social and affective behaviors are perceived as more natural and effective when people interact with them (Heerink et al., Citation2010; Kanda et al., Citation2004; Nass et al., Citation1995; Tsiourti, Citation2018; Young et al., Citation2008). Further, these kinds of behavior bring an increased sense of credibility, reliability, and authenticity of the system (Becker et al., Citation2007). Krämer et al. (Citation2011) address the aspect of sociability and especially companionship within CAs as “[…] it becomes more and more feasible that artificial entities like robots or agents will soon be parts of our daily lives.” (Krämer et al., Citation2011, p. 474').

Today, there are already several CAs with companionship characteristics aimed at building long-term relationships with their users. Replika,Footnote2 for example, is an AI companion, as the providers themselves call it, which specializes in extended conversations and long-term relationships to provide its users with an omnipresent friend. Microsoft has developed XiaoIce,Footnote3 a service they call an AI companion that aims to establish an emotional connection with its users in order “[…] to satisfy the human need for communication, affection, and social belonging.” (Zhou et al., Citation2020, 53'). Although both systems are very successful (Replika has 7 million users and XiaoIce has around 660 million users), the design decisions made to build a long-term relationship have not been studied and scientifically derived (Zhou et al., Citation2020). There is already research (Bickmore & Picard, Citation2005; Vardoulakis et al., Citation2012) investigating the use of companionship in CAs to establish a long-term relationship with their users. However, prescriptive knowledge, which is transferable to different instances, is especially required by designers of different CAs developed to incorporate aspects of companionship (Diederich et al., Citation2022; McTear, Citation2018). Prescriptive design knowledge, such as design principles and design theories, add value by providing recommendations to all stakeholders involved in designing and developing CA instantiations (Gregor & Jones, Citation2007; Gregor et al., Citation2020). With theory, and design theory specifically, one can make distinctions and order how we understand phenomena, form interpretations and explanations, and develop informative accounts for researchers and practitioners (Cornelissen et al., Citation2021). Hence, the primary objective of this paper is to address our lacking insight by constructing a design theory for so-called virtual companionship (VCS) that can help to build CAs equipped with aspects of companionship, where companionship is identified as a beneficial added value. With VCS, the human-machine interaction can be enhanced, because it encourages longer-lasting and more intuitive conversations (Bickmore & Picard, Citation2005; Gnewuch et al., Citation2017; Seymour et al., Citation2018), increases the likability and interpersonal trust users have in a system (Bickmore & Picard, Citation2005), and builds a foundation to enable collaborative scenarios between humans and machines (Li et al., Citation2013; Seeber et al., Citation2019). In building a theoretical framework for VCS, we first turn to basic human-human interaction theories and discuss their applicability to human-machine interaction, as these provide us with a starting point to explain VCS (Krämer et al., Citation2011, Citation2015). After all, humans inevitably exhibit human patterns in their communication with machines (Nass & Moon, Citation2000; Nass et al., Citation1994). Following Krämer et al.’s (Citation2011) prior work on VCS, however, we do not see designing virtual companions to replicate humans as an ideal, since machines that are too human-like can have negative perceptual effects (Mori, Citation2012; Seymour et al., Citation2021). Rather, we focus on integrating individual social elements on a well-balanced level to enrich the interaction with CAs (Feine et al., Citation2019; Seeger et al., Citation2018).

Building on these insights, our research pursues four objectives, namely (1) to derive theoretically grounded design principles for VCS that can be applied to CAs, summarized as a nascent design theory (Gregor & Hevner, Citation2013; Gregor & Jones, Citation2007), (2) to instantiate the theoretically grounded design principles in the form of artifacts and to evaluate these artifacts, (3) to analyze existing CAs that incorporate companionship characteristics, to determine the suitability of our proposed design principles, and (4) to derive design features that guide the translation of the proposed design principles into instantiated artifacts.

The nascent design theory thus serves the purpose of creating design knowledge, and aims to address our overarching design-oriented research question (RQ) (Thuan et al., Citation2019):

RQ: How is virtual companionship made manifest and how should it be designed?

To address this research question, we define two more narrowly focused sub-questions (SQs) that address the aspect of how virtual companionship should be designed and how it makes itself manifest:

SQ1: Which requirements, design principles, and design features establish virtual companionship?

SQ2: How do CAs demonstrate companionship elements and how do users perceive them?

In answering the two SQs, we find answers to the overarching RQ and contribute to constructing our nascent design theory. We followed a systematic and iterative design science research (DSR) approach (Hevner et al., Citation2004) in developing our VCS design theory, relying on the adapted process model proposed by Kuechler & Vaishnavi (Citation2008) , which emphasizes the contribution and knowledge generation of DSR. After theoretically deriving the design principles for VCS, we developed an instantiation (Sarah) which we then evaluated in a within-subject design experiment. Since design principles are prescriptive statements in a means-end relationship, an experimental study is necessary to empirically demonstrate the design decisions’ effects in evaluating the prescriptive statements (Gregor & Jones, Citation2007; Gregor et al., Citation2020). In a second study, we examined Replika’s features and functionality based on the previously derived principles for VCS design. Next, eight design experts provided feedback on the extent to which Replika meets the design principles we established and which design features are used. This mixed-method approach combines quantitative and qualitative research methods to “develop rich insights into various phenomena of interest that cannot be fully understood using only a quantitative or a qualitative method” (Venkatesh et al., Citation2013, p. 22). Thus, this approach enables the evaluation of the design principles and the development of our nascent design theory. Further, with this multi-method approach, we aimed to evaluate both, a self-designed virtual companion (Sarah) and an existing solution (Replika). Thus, the limitations of the two approaches (Sarah’s potential low external validity and Replika’s potential low internal validity) can be weighed and the results can be properly combined. Sarah is a designed expository instantiation following Gregor and Jones (Citation2007) that shows how the design principles can be instantiated. Furthermore, causal statements based on the empirical findings regarding the means-end relationship of our design principles can be made through the within-subject experiment. Through the design experts’ evaluation of Replika, our prescriptions can be further extended with concrete design features. Our nascent design theory eventually contributes to a better understanding of an emerging phenomenon (i.e., VCS) by conceptualizing, describing, explaining, and informing us on characteristics, distinctions, as well as design decisions and their impact and antecedents related to VCS (Gregor, Citation2006; Sandberg & Alvesson, Citation2021).

2. Theoretical foundations

2.1. Conversational agents

CAs are systems that allow their users to interact with them using natural language (Gnewuch et al., Citation2017; McTear et al., Citation2016). The development of CAs started in the 1960s with, for example, the chatbot ELIZA (Weizenbaum, Citation1966), one of the first text-based CAs (Shah et al., Citation2016). While in the past chatbots worked using simple pattern matching, chatbots today are more intelligent due to developments in the fields of machine learning and natural language processing (Knijnenburg & Willemsen, Citation2016; McTear et al., Citation2016; Shawar & Atwell, Citation2007). Public interest in CAs especially grew after Apple’s Siri and later Amazon’s Alexa or the Google Assistant were introduced, since these brought virtual assistants to the smartphones and homes of large numbers of people (Diederich et al., Citation2019a; Maedche et al., Citation2016; McTear et al., Citation2016). One of the differences between chatbots and virtual assistants is the input/output modality. While chatbots are mostly text-based, virtual assistants usually communicate using spoken language (Gnewuch et al., Citation2018b). Chatbots are often used for domain-specific purposes, for example, in giving customer support, and are mostly used via text-based input (Diederich et al., Citation2019b; Gnewuch et al., Citation2017). There are also interactive voice response interfaces which are chatbot applications that enable interaction via speech and are used, for example, in telephone hotlines (Clark et al., Citation2019a). The technological progress through AI has made it possible to interact with a CA using natural language and not only a command-based language that uses only predefined commands (Diederich et al., Citation2022; Gnewuch et al., Citation2017; Rzepka & Berger, Citation2018).

These developments enabled virtual assistants such as Siri and Alexa, whose primary function is to assist their users (Gnewuch et al., Citation2017; McTear, Citation2017). Virtual assistants are often used with speech, even if not exclusively so (Gnewuch et al., Citation2017). Virtual assistants differ from chatbots in a general-purpose approach; for example, they assist users in performing daily tasks (Gnewuch et al., Citation2017; Guzman, Citation2017). Illustrative current use cases are smartphones that execute functions such as creating calendar entries, sending messages, or asking for the weather forecast. Thus, virtual assistants offer a new way of interacting with information systems (Morana et al., Citation2017). However, as stated before, they do so only by providing personal assistant functionalities, without social behaviors being included in the actual interaction with the system (Gnewuch et al., Citation2017; Krämer et al., Citation2011; Rzepka & Berger, Citation2018). below gives an overview of the distinct terms.

Table 1. Relevant definitions.

Gnewuch et al.’s (Citation2017) definition of a virtual assistant, given above, restricts CAs to their pure assistance services, e.g., giving support in entering calendar entries or navigating web pages (McTear et al., Citation2016). However, recent research shows that machines are capable of providing more than assistance services, especially in building interpersonal relationships. Several studies have shown that machines are able to build relationships with humans, for example, by giving CAs feelings of trust and establishing a long-term relationship with them (Benbasat & Wang, Citation2005; Hildebrand & Bergner, Citation2021; Krämer et al., Citation2011; Lee et al., Citation2021). Thus, the evolution from virtual assistant to virtual companion is particularly notable in the fact that the CA focuses on developing a relationship with its user.

The evolution of CAs from being simple virtual assistants or chatbots to virtual companions designed to have a benevolent and long-term relationship with humans, can be explained primarily by the fact that humans often mindlessly apply the same social heuristics used in human interaction to the CA when they display characteristics similar to those of humans (Moon, Citation2000). The evolutionary steps in the development from CA to virtual assistant to virtual companion are illustrated in . Notably, chatbots and virtual assistants can be enhanced by adding characteristics and principles of the VCS.

Figure 1. From assistance to companionship.

Figure 1. From assistance to companionship.

2.2. Computers are social actors

Nass and Moon’s (Citation2000) work provides the foundation for reflection on the social behavior of CAs. They demonstrated that humans apply social rules in their interaction with machines, and that machines can incorporate such social behavior. For example, in an experiment involving a computer tutor with a human voice, Nass et al. (Citation1994) found that users respond to different voices produced by the same computer as if they were distinct social actors and to the same voice as if it were the same social actor, regardless of whether the voice came from the same or a different computer. These findings prompted the formation of the “Computers are Social Actors” (CASA) paradigm, also referred to as the “social response theory” (Nass & Moon, Citation2000; Nass et al., Citation1994). Further findings have established that if the machine has features normally associated with humans, such as interactivity, the use of natural language, or human-like appearance, users respond with social behavior and social attributions (Moon, Citation2000). Since such human-like design can amplify humans’ social responses, interactions between individuals and computers have the possibility of being social (Feine et al., Citation2019; Nass & Moon, Citation2000; Nass et al., Citation1994). Humans do this because they are accustomed to only humans showing social behavior. Notably, the CASA paradigm does not claim that users regard machines as human beings. Nass and Moon (Citation2000) emphasize that they are aware the interlocutors are machines. Even so, they behave socially toward them, thus indicating “a response to an entity as a human while knowing that the entity does not warrant human treatment or attribution” (Nass & Moon, Citation2000, p. 94). Such behavior, referred to as ethopoeia, was also witnessed in that people showed politeness to machines so that the interaction with them closely resembled an equal relationship in which reciprocity was evident (Moon, Citation2000; Nass & Moon, Citation2000). If a computer acts socially, humans mindlessly apply known social rules in the interaction with the computer as well (Fogg, Citation2002; Nass & Moon, Citation2000). Thus, elements of social conventions that typically guide interpersonal behavior can also be applied to human-machine interaction (Moon, Citation2000).

Although Nass and Moon (Citation2000) refer to human response to machines’ actions in general, it has since been determined that the CASA paradigm is also applicable to CAs in particular (Porcheron et al., Citation2018; Reeves et al., Citation2018). In the case of speech-based CAs such as Amazon’s Echo, their ubiquity in the household even leads to family members competing to control the Echo during dinner, illustrating how the technology triggers social behavior in users (Porcheron et al., Citation2018). Diederich et al. (Citation2022). This emphasizes the importance of the CASA theory as one of the most significant theories in the context of CA research. Social reactions to a CA’s behavior can be enhanced if the CA exhibits human-like behavior in giving social cues (Feine et al., Citation2019; Seeger et al., Citation2018). Many users perceive the integration of such human-like elements as pleasant (Feine et al., Citation2019). For example, Gong’s (Citation2008) results show that the more human-like the CA, the more naturally users responded to the CA’s interaction behavior. Further, integrating social cues can encourage the establishment of a trusting relationship (Bickmore & Picard, Citation2005; De Visser et al., Citation2016), which leads to CAs being seen as reliable and composed. Users particularly trusted the CA’s opinions when they had to take important decisions (Gong, Citation2008). Using Amazon’s Alexa illustratively, Doyle et al. (Citation2019) define dimensions for implementing human-likeness in speech-based CAs, which include establishing a social bond with the user or utilizing human-like linguistic content in communication. This highlights the feasibility of anthropomorphic CAs. Still, we need to note that significant differences remain between how humans communicate with one another and how humans and CAs communicate (Doyle et al., Citation2019). Furthermore, human-like design features should be used thoughtfully in CAs (Diederich et al., Citation2022; Seeger et al., Citation2018), as anthropomorphic design can also be perceived negatively (Doyle et al., Citation2021, Citation2019; Mori, Citation2012). A number of studies emphasize the CASA paradigm’s importance and apply the theory to a variety of different CAs with a view to designing a human-machine relationship (Kim et al., Citation2013; Lee & Choi, Citation2017; Pfeuffer et al., Citation2019b; Qiu & Benbasat, Citation2010). Thus, we find the CASA paradigm appropriate to gain an understanding of how people interact with CAs and how relationships can be built between humans and CAs.

2.3. Theories of interpersonal relationship

Humans have a fundamental need to develop long-term interpersonal relationships (Baumeister & Leary, Citation1995; Cacioppo & Patrick, Citation2008; Deci & Ryan, Citation1985; Krämer et al., Citation2012). Long-term studies such as Bickmore and Picard (Citation2005) show that CAs are capable of building relationships and thus satisfying this need. Such studies then demonstrate increased perception of these interactions. An interpersonal relationship is an advanced and close connection between one or more persons that can take on different forms, such as friendship, marriage, neighborliness, or companionship (Kelley & Thibaut, Citation1978). As an important and pervasive human need, meaningful interpersonal relationships arise over time and exist when the probable course of future interactions differs from such courses between strangers (Hinde, Citation1979). Depending on the nature of the relationship, a number of theories have been established in relationship science, i.e., in the study of interpersonal relationships (Berscheid, Citation1999). This field has long been well researched. We consider selected theories of this kind that are most relevant to companionship (Krämer et al., Citation2011, Citation2012; Skjuve et al., Citation2021), i.e., the need for belonging (Baumeister & Leary, Citation1995), the social exchange theory (Blau, Citation1968; Homans, Citation1958), the social penetration theory (Altman & Taylor, Citation1973), the equity theory (Walster et al., Citation1978), the common ground theory (Clark, Citation1992), the theory of mind (Carruthers & Smith, Citation1996; Premack & Woodruff, Citation1978), and interpersonal trust theory (Rotter, Citation1980).

As social beings, humans have a strong desire for interpersonal relationships motivated by the need to belong (Baumeister & Leary, Citation1995). This leads to people’s interest in building and maintaining positive and warm relationships in their pursuit of happiness (Baumeister & Leary, Citation1995; Krämer et al., Citation2012). Building and maintaining functioning relationships involves reciprocal behavior, in which costs and rewards are balanced (Blau, Citation1968; Homans, Citation1958). This social exchange drives relationship decisions, establishes a norm between the partners, and subsequently maintains and stabilizes the relationship (Blau, Citation1968; Homans, Citation1958; Krämer et al., Citation2011). Further, interpersonal relationships are characterized by depth and breadth of self-disclosure, which is also referred to as the social penetration theory (Altman & Taylor, Citation1973; Carpenter & Greene, Citation2015). According to Skjuve et al. (Citation2021), consideration of the social exchange and social penetration theories are crucial to build a relationship with a CA since, in the long run, the phenomena they propagate help to bond users with the CA, as existing applications like Replika have demonstrated (Skjuve et al., Citation2021). This equitable input-outcome ratio eventually leads to a co-equal relationship, as the equity theory predicts (Sprecher, Citation1986; Walster et al., Citation1978). Equity contributes to partners feeling valued and satisfied in an interpersonal relationship, which is also relevant for the relationship between humans and machines (Fox and Gambino Citation2021). The formation of interpersonal relationships requires developing common ground over time (Clark, Citation1992; Krämer et al., Citation2011), building the ability to grasp and share the mental perspective of the relationship partner and to predict their actions. This phenomenon is often described as people being “soulmates” or “mindreaders” and defines deep interpersonal relationships that are especially relevant in communication between human interactants (Krämer et al., Citation2011). Common ground is created in interaction, for example, when conversation partners make an effort to find a common communication style (Clark & Brennan, Citation1991; Clark & Schaefer, Citation1987), ensure mutual understanding (Clark, Citation1996; Svennevig, Citation2000), or draw on shared knowledge and assumptions they have about each other (De Angeli et al., Citation2001). With the abovementioned aspects incorporated, CAs are able to form a common ground when they communicate and interact with the user (Bickmore et al., Citation2018; Chaves & Gerosa, Citation2021). Interpersonal trust, which is an indispensable condition in relationships, strongly influences the smooth functioning of such relationships. Rotter (Citation1980) defines interpersonal trust as the willingness to be vulnerable to the actions of a relationship partner. As mentioned, several researchers have already shown that CAs and humans are capable of interacting in a trusting relationship (Benbasat & Wang, Citation2005; Hildebrand & Bergner, Citation2021; Lee et al., Citation2021). This then leads to individual well-being, better communication and interaction overall, and finally to a benevolent relationship (Rotter, Citation1980; Schroeder & Schroeder, Citation2018).

2.4. Derivation of the virtual companion

That humans can feel socially related to an individual whose existence is of an artificial nature, is not a new idea: Horton and Wohl (Citation1956) originally recognized parasocial interaction, which describes the non-reciprocal manner in which an audience member interacts with a media persona. Parasocial interaction can become a parasocial relationship, which is defined as the ongoing process of affective and cognitive responses to a media persona outside the viewing time, and leading to a long-term relationship being established (Stever, Citation2017). Although the original concepts of parasocial interaction and parasocial relationship referred mainly to media personas on television as an individual’s relationship partner, research has shown that these concepts can also be applied to human-machine interaction (Krämer et al., Citation2012). Such application of interpersonal relationship theories to human-machine interactions, enables the crafting of a new kind of relationship between human and machine, namely VCS. Earlier research already followed similar approaches, such as using the artificial companion concept (Danilava et al., Citation2012; Wilks, Citation2005), the relational agent (Bickmore & Picard, Citation2005), or Seymour et al.’s (Citation2018) suggestion of building relationships with visual cognitive agents. These relationships are more a side-product of their main research interest in using natural face technology to create a realistic visual presence. Bickmore and Picard (Citation2005) successfully demonstrated the relevance and benefits of long-term human-machine relationships in a study on an exercise adoption system. Danilava et al. (Citation2012) relied on a concept Wilks (Citation2005) introduced, in a first attempt at requirement analysis, and Krämer et al. (Citation2011) built a solid foundation for understanding human-CA relationships with their theoretical framework proposed for designing artificial companions by applying social science theories to human-machine interaction. The authors discuss aspects of sociability in developing a theoretical framework for sociable companions in human-artifact interaction. The identified theories were grouped into a micro-, meso- and macro-level according to their level of sociability in order to guide, among other things, the “implementation of the companions’ behaviors” (Krämer et al., Citation2011, p. 475). A fundamental focus of their work is to subdivide interpersonal theories according to aspects of sociability, and thereby to narrow down the definition of companionship. Further, they raised several questions on the usefulness of their approach and its benefits in deriving guidelines. We find that their theory of companions provides a solid foundation introducing a variety of important theories upon which we can build our VCS research.

The research presented above, that draws on the CASA paradigm, suggests that various aspects of human-to-human relationships and human-to-human interactions are transferable to the relationship between humans and CAs. Although several studies already show the evolution from virtual assistants to virtual companions, we lack theoretically well-founded design knowledge and theories to simplify the virtual companion designing process (Clark et al., Citation2019a; Krämer et al., Citation2011; McTear, Citation2018). This is reflected in the generalizable design knowledge being inadequately available and in multiple user contexts rarely being analyzed (Clark et al., Citation2019a). Therefore, we require prescriptive knowledge, which is transferable to different instances (Diederich et al., Citation2022). Such design principles and theories add value by providing prescriptive recommendations to the parties involved in designing and developing CA instantiations (Gregor & Jones, Citation2007; Gregor et al., Citation2020). Additionally, different users perceive virtual companions very differently, which can have positive as well as negative effects, so one-size-fits-all solutions are inappropriate (Krämer et al., Citation2011). Thus, the consideration of individual user needs must also be reflected in prescriptive design knowledge (Dautenhahn, Citation2004; Krämer et al., Citation2011). Due to the user-specific individuality of companionship, design decisions should not be derived from theory alone; individual choices should be left to the user, for example, through the possibility of the CA being customizable (Krämer et al., Citation2011). According to recognized approaches to crafting and presenting prescriptive knowledge in DSR (Gregor & Hevner, Citation2013; Kuechler & Vaishnavi, Citation2008), design knowledge for VCS needs further elaboration to effectively support the instantiation of virtual companion artifacts.

3. Methodology

In developing our nascent design theory for VCS, we followed a systematic and iterative DSR approach. DSR combines characteristics of behavioral science and design science ensuring that artifacts are designed (construction) based on business needs and requirements (relevance) and are built on relevant knowledge gained from theories, frameworks, and methods (rigor; Hevner et al., Citation2004). The nascent design theory is successively and iteratively developed following the adapted process model Kuechler and Vaishnavi (Citation2008) proposed as in , and that emphasizes the contribution and knowledge generation of DSR. This continuous development ensures a rigorous process to create our nascent design theory.

Figure 2. The DSR approach based on Kuechler and Vaishnavi (Citation2008)

Figure 2. The DSR approach based on Kuechler and Vaishnavi (Citation2008)

Kuechler and Vaishnavi (Citation2008) argue that within a design cycle their five steps can generate various contributions. Constraint knowledge (circumscription) can already arise in the development and evaluation steps, while design knowledge emerges after the evaluation and conclusion, all contributing to the body of knowledge. The constraint knowledge emerges through the application of theories, instantiations of technology, and the analysis of contradictions that arise (Kuechler & Vaishnavi, Citation2008). Consequently, a design cycle can have several DSR outcomes in the different process steps. The levels of artifact abstraction in DSR (see, ) vary, ranging from models, processes, instantiations, methods, or software. Gregor and Hevner (Citation2013) note that any artifact has a specific level of abstraction, while proposing three main levels for categorization. An artifact can be described by the two attributes “abstract” and “specific” and by their general knowledge maturity. A specific and limited artifact (level one), such as software products or implemented processes, contribute situated and context-specific knowledge. An abstract and complete artifact (level three), such as design theories about embedded phenomena, contribute broad and mature kinds of knowledge. Design principles and constructs, methods, models, or design features (level two) lie between a solely abstract and a specific artifact, contributing nascent design theories or operational principles (Gregor & Hevner, Citation2013). Hence, there can be different degrees of abstraction depending on the extent of the artifact’s generalization in artifact level 2.

Figure 3. Levels of artifact abstraction.

Figure 3. Levels of artifact abstraction.

Recently, the call for a theoretical contribution has been emphasized (Gregor & Hevner, Citation2013; Kuechler & Vaishnavi, Citation2012). Gregor and Hevner (Citation2013) argue that summarizing the prescriptive knowledge in the form of a nascent design theory, can substantially increase the knowledge contribution and impact. Hence, we follow and use the eight components of a design theory introduced by Gregor and Jones (Citation2007), which are: 1. Purpose and scope (set of meta-requirements, which specify what the system is for), 2. Constructs (entities of interest), 3. Principles of form and function (the meta-design in the form of design principles), 4. Artifact mutability, 5. Testable propositions (truth statements), 6. Justificatory knowledge (underlying theory), 7. Principles of implementation (description for implementation in the form of design features) and 8. Expository instantiation (physical implementation of the artifact).

As our understanding of the individual components has recently evolved, we build on current methodological findings. The principles of form and function, for example, are understood as design principles according to Gregor et al. (Citation2020), and the meta-requirements are considered as the artifact related requirements that Walls et al. (Citation1992) originally presented, on which Gregor and Jones (Citation2007) built (Iivari, Citation2019). Following Möller et al. (Citation2020) and Meth et al. (Citation2015), we formulated the principles for implementation in the form of design features, which show how the design principle prescriptions can be physically instantiated, while at the same time leaving space for their technological implementation, since they are more generally formulated.

Gregor and Jones (Citation2007) argue that design theories manifest through the building and evaluation of various instantiations. These instantiations should go beyond prototypical artifacts that are tested in laboratory settings to show real world application. We address this by including two different instantiations of our nascent design theory, which are Sarah, a self-designed virtual companion and Replika an existing real work application. Kuechler and Vaishnavi (Citation2008) emphasize the necessity of a rigorous evaluation to codify design knowledge as design theories because evaluation findings extend, refine, confirm, or reject the constructed design knowledge. To conduct this evaluation as holistically as possible, we follow a mixed-method approach and the guidelines of Venkatesh et al. (Citation2013) on combining quantitative and qualitative methods. In this way, we could discover and develop integrative insights based on quantitative and qualitative findings. The first guideline requires checking the appropriateness of a mixed-method approach, which is advised if researchers need to “gain a better understanding of the meaning and significance of a phenomenon of interest” (Venkatesh et al., Citation2013, p. 37). Combining quantitative and qualitative analysis strengthens our understanding of CAs, as does the independent instantiation (Sarah) and the analysis of an existing instantiation (Replika). The second guideline determines the strategy, which in our case is sequential since we designed the two studies consecutively to build on one another. The final guideline Venkatesh et al. (Citation2013) propose is that meta-inferences should be developed as theoretical statements which serve to combine the quantitative and qualitative methods’ findings and to see them as a holistic explanation of the phenomenon. This theoretical statement, explaining the results of the mixed-method approach is the nascent design theory of VCS.

Our DSR project outcomes are listed in and presented in the following sections. This paper’s main contribution is our nascent design theory for VCS, which summarizes the design knowledge collected thus far. The instantiation of the virtual classmate Sarah was conceptually designed, developed as a prototype, and evaluated. The real-world virtual companion Replika was then analyzed as an existing instantiation. This paper presents Sarah and Replika as expository instantiations. According to Gregor and Hevner’s (Citation2013) three levels of DSR contribution types our instantiation, Sarah, is a level 1 artifact and the design principles, as well as the nascent design theory for VCS are a level 2 artifact.

By summarizing our results and contributions in the form of a nascent design theory, we aim to extend the body of knowledge for researchers and practitioners, and provide an initial step toward a comprehensive design theory (Gregor & Jones, Citation2007).

4. Designing virtual companionship

4.1. Deriving design principles for virtual companionship

According to Gregor and Jones (Citation2007), meta-requirements are derived to serve a whole class of artifacts, rather than a single instance of a system. Additionally, they are part of a design theory and define its purpose and scope (Gregor & Jones, Citation2007). Meta-requirement elicitation is a process that involves multiple sources, such as literature reviews and core theories (Möller et al., Citation2020). Our literature review was conducted in a semi-structured systematic manner (Webster & Watson, Citation2002). For this purpose, we used various search queries intended on the one hand for studies about mapping human-like relationships to human-machine interaction, and on the other hand for gaining insight around CA design in general. The results and findings were independently collected and aggregated in an Excel spreadsheet by three researchers in the author’s team. The final table was divided into the three main areas: human-human relationship, human-machine interaction, and the VCS to bridge the gap. Based on this table, the three researchers individually derived the meta-requirements, which they then collaboratively aggregated, resulting in 25 meta-requirements (MR1 to MR25) for VCS. To adhere to the meta-requirements, we translated them into five design principles (DP1 to DP5) as proposed by Möller et al. (Citation2020), who state that design principles are formulated “[…] as a response to meta-requirements.” (p. 212). As the design principles’ general objective, we followed Chandra et al.’s (Citation2015) suggestion, later adapted by Gregor et al. (Citation2020) for action and materiality oriented design principles, which “[…] prescribe what an artifact should enable users to do and how it should be built in order to do so.” (Chandra et al., Citation2015, p. 4043). Here, the authors point out the importance of involving stakeholders (Purao et al., Citation2020) in formulating design principles with the goal that they will be “prescriptive statements that show how to do something to achieve a goal” (Gregor et al., Citation2020, p. 1622'). Gregor et al. (Citation2020) analyzed existing formulations of design principles, identified the most important components of a design principle, and proposed the anatomy of a design principle, stating that a design principle consists of the goal, the implementer and user, the context, the mechanism, and the rationale. We do not additionally list the context since it is the same in all five of our proposed design principles, i.e., the interaction between human and virtual companions. The set of design principles has been iterated and refined throughout this project, following the iterative approach of a design process (Kuechler & Vaishnavi, Citation2008). We present the final design principles in the following sections.

According to the social response theory (i.e., CASA paradigm) users attribute human characteristics to non-humans and invest human-like characteristics in their designed behavior (Moon, Citation2000; Nass & Moon, Citation2000), which leads to an adapted interaction with machines that are fundamentally social (Nass & Moon, Citation2000). Humans tend to form and appreciate companionships as social bonds with other humans (Fehr, Citation1996; Hinde, Citation1979; Kelley & Thibaut, Citation1978). If we create VCS between a human and a machine, the machine should also be designed with a human-like appearance and behavior, which is possible according to the social response theory (MR1). A human-like design usually integrates so-called social cues in the CA design (Feine et al., Citation2019; Seeger et al., Citation2018). Such cues are design features originally associated with humans, so that the machine has features such as the ability to use human gestures, tell jokes, or speak with a human-like voice (Feine et al., Citation2019). Integrating social cues into intelligent systems such as CAs consistent with the CASA paradigm, encourages people to act socially toward them (Feine et al., Citation2019; Fogg, Citation2002; Nass & Moon, Citation2000). This yields diverse benefits. For example, it has been shown that incorporating social cues in CAs can increase trust and perceptions of credibility toward the CA (Bickmore et al., Citation2005; Demeure et al., Citation2011; De Visser et al., Citation2016). Further, such integration fosters the establishment of a social bond between the human and the CA (Bickmore & Picard, Citation2005), which opens the potential for VCS (Krämer et al., Citation2011). According to Vardoulakis et al. (Citation2012), CAs designed to build a trustful relationship are “typically designed as computer-animated, humanoid agents that simulate face-to-face dialogue with their users” (Vardoulakis et al., Citation2012, p. 290). However, it is important to consider current standards, norms, and values of the society to ensure the responsible and accepted use of trusted forms and applications of AI (e.g., degree of autonomy, independence, or human-likeliness; Abbass, Citation2019)(MR2). With technological progress, over time the socially accepted standards are also changing. Whereas in the past people would not have spoken to their smartphones, we already find this behavior and we expect it to be common in the future. Recent research has shown that the uncanny valley theory (Mori, Citation2012) is also applicable to CAs (Gnewuch et al., Citation2018a; Seeger et al., Citation2018; Seymour et al., Citation2019). This should then be taken into account in designing a virtual companion (MR3). The theory argues that individuals have a greater affinity to machines that are more realistic, until the machine becomes so human-like that users find its non-human imperfections disturbing. This is the point at which the affinity dramatically drops and the CA is experienced as uncanny. The uncanny valley effect relates not only to the appearance of the virtual companion, but also its verbal interaction. Thus, users of speech-based CAs can even perceive voices that are too human-like as unpleasant (Clark et al., Citation2021; Doyle et al., Citation2021, Citation2019). Mori (Citation2012) suggests that with a very high degree of realism, and thus anthropomorphism, the uncanny valley effect can be overcome. According to this, affinity would reach its highest level, when the human imitation can no longer be distinguished from a real human. We emphasize that we refer to CAs which usually interact with their users via a messaging platform or a speech interface. These have to be distinguished from embodied CAs (Cassell et al., Citation1999), which refer to virtual representations of humans, including a physical appearance with bodily gestures and facial expressions, which would have more profound requirements (Seeger et al., Citation2018). According to Seeger et al. (Citation2018), who developed a framework for the design of anthropomorphic CAs, it is important not to follow a “more is more” approach when designing a CA’s appearance and behavior, but rather to balance human identity (e.g., a human-like representation or demographic information like gender or age), non-verbal features (e.g., hand gestures, facial expressions, or the use of emojis) and verbal characteristics (e.g., the choice of words and sentences) (MR4). Qiu and Benbasat (Citation2009), having investigated the construct of social presence referring to the feeling of being with another, described a quasi-social relationship between human and machine, which can be used to measure the user’s perception of an agent’s social characteristics. By incorporating a human-like appearance and behavior the experience of social presence can be increased (Qiu & Benbasat, Citation2009). Thus, we propose DP1 in .

Table 2. Principle of human-like design.

Due to the high individuality of the concept of companionship, it is difficult to derive design decisions based solely on theory (Krämer et al., Citation2011). Every person has their own ideas and requirements for a companionship relation to be satisfactory. This is dependent on individual characteristics and behavior that appeals to the user (Bukowski et al., Citation1993; Krämer et al., Citation2011) (MR5). Especially when it comes to an embodied virtual companion’s appearance, aspects such as what gender the system should have, whether it should have a gender at all, what the face and hair color, but also the clothing should look like, are very individual (Siemon & Jusmann, Citation2021; Xiao et al., Citation2007). Several studies have investigated the bias in and implications of gender decisions in CA design, yet none draw a final conclusion, especially for VCS (Feine et al., Citation2020; Pfeuffer et al., Citation2019a; Winkle et al., Citation2021). Therefore, we suggest letting the users decide about the preferred gender of their virtual companion in the particular moment (Feine et al., Citation2020) (MR6). This would mean that the user is given the opportunity to customize aspects of the virtual companion’s appearance and demeanor. To create an artifact that is tailored to the user’s preferences (MR7), it is thus necessary to enable adaptivity and adaptability to let the user customize certain features.

To understand the behavior and language characteristics of the user, it is important to get to know him or her (Danilava et al., Citation2012; Seymour et al., Citation2018) (MR8). Understanding the user better over time, steadily improves the ability to adapt to him or her (Park et al., Citation2012) (MR9). This ability is important. Not only to maximize a better interaction, but also to help establish common ground and mutual language (Danilava et al., Citation2012; Krämer et al., Citation2011; Turkle, Citation2010). When interpersonal relationships are formed, common ground is necessary (Clark, Citation1992; Krämer et al., Citation2011) (MR10), which enables the ability to better take the mental perspective of the relationship partner and predict his or her actions (Carruthers & Smith, Citation1996; Premack & Woodruff, Citation1978). The virtual companion should therefore use a language that is familiar to the user (Park et al., Citation2012) (MR11) and must be designed to understand the user’s language as well as possible (MR12), as this is an inevitable condition for satisfactory communication (Gnewuch et al., Citation2017). Additionally, the rule of attraction, according to which the machine adapts to the user can enhance the overall communication satisfaction (Hecht, Citation1978; Nass et al., Citation1995; Park et al., Citation2012). Thus, we propose DP2 in .

Table 3. Principle of adaptivity and adaptability.

social exchange theory assumes that relationships are driven by rewards and costs, where costs are elements with negative value for the individual, such as effort put into building the relationship (Blau, Citation1968; Homans, Citation1958). In contrast, rewards are elements with a positive value, such as acceptance or support (Blau, Citation1968; Homans, Citation1958). Applied to a human-machine relationship the costs to the user could be releasing personal data or paying attention, and the rewards delivered by the machine would be support or information of high quality. For a stable relationship, cost and rewards should be balanced in the VCS (Blau, Citation1968; Krämer et al., Citation2012) (MR13). Therefore, the virtual companion should reward the user when costs are invested (MR14) and it should incorporate a reciprocal exchange (MR15) in order to build an equal relationship (Blau, Citation1968; Gouldner, Citation1960; Krämer et al., Citation2012). Collaborative tasks could be given to the user and the virtual companion to reward the user in fulfilling the task and to encourage a reciprocal exchange (Kapp, Citation2012). Moreover, the virtual companion has to act proactively and reactively by, for example, demanding and asking, as well as giving and answering in an equal measure (L’Abbate et al., Citation2005) (MR16). Thus, we propose DP3 in .

Table 4. Principle of proactivity and reciprocity.

This reciprocal exchange of information, however, must be governed by a set of guidelines that protect the user’s privacy. As in a close relationship, user information should be kept confidential and not disclosed to third parties (Elson et al., Citation2018; Solove, Citation2004) (MR17). The user should also be able to determine what he or she wants to disclose and what information is necessary to obtain which value (Fischer-Hübner et al., Citation2016; Schroeder & Schroeder, Citation2018; Taddei & Contena, Citation2013) (MR18). The handling of this information must therefore be transparent so that it is clear to the user how it is stored and what added value it creates (Saffarizadeh et al., Citation2017; Turilli & Floridi, Citation2009; Wambsganss et al., Citation2021) (MR19). Another important aspect in the AI design of is to observe ethics (Wambsganss et al., Citation2021; Yu et al., Citation2018). Due to an abundance of different perspectives and ethical norms, such as machine ethics (Anderson & Anderson, Citation2011) or robo-ethics (Veruggio et al., Citation2016), developers should decide on and adhere to suitable ethical commitment, and also design the virtual companion as an entity that adheres to ethical principles (MR20). Incorporating transparent treatment of the users’ information, preserving privacy, and implementing ethical rules can increase the users interpersonal trust in the virtual companion (Al-Natour et al., Citation2010; Fischer-Hübner et al., Citation2016; Saffarizadeh et al., Citation2017; Schroeder & Schroeder, Citation2018; Turilli & Floridi, Citation2009). Thus, we propose DP4 in .

Table 5. Principle of transparency, privacy and ethics.

People are naturally inclined to form social attachments, as “human beings are fundamentally and pervasively motivated by a need to belong, that is, by a strong desire to form and maintain enduring inter-personal attachments” (Baumeister & Leary, Citation1995, p. 522). This need to belong is what motivates humans to seek the company of others (Krämer et al., Citation2012) (MR21). A virtual companion, designed according to the needs and preferences of its user would give the user the impression of having someone familiar and likeable spending time with him or her, and would therefore enable the feeling of being together, rather than being alone (Krämer et al., Citation2011; Qiu & Benbasat, Citation2009; Seymour et al., Citation2018). This feeling of being with another is essential for companionship, since being with someone is the central part of our companionship definition. Therefore, this kind of feeling should also be triggered in the user, who interacts with the virtual companion as in a human-human relationship, to reach VCS (MR22). Due to its artificial nature, the virtual companion can and should always be available to its user. Humans as social beings seek the company of other people regularly (Baumeister & Leary, Citation1995). Behind purely being in the company of the other, however, is the fundamental desire to build close, positive, and warm relationships that result in the motivation to build friendships (Krämer et al., Citation2011). Hence, companionship can be seen as a part of friendship (Bukowski et al., Citation1994; Krämer et al., Citation2011; Mendelson & Aboud, Citation2012), thus an overall goal of the virtual companion could be oriented toward principles of friendship (Fehr, Citation1996; Rawlins, Citation2017) in order to make and maintain friendship (MR23). As VCS can be seen as a form of human-machine collaboration (Li et al., Citation2013), principles of collaboration (Siemon et al., Citation2017) can also be taken into account. Such collaboration is defined as the joint effort toward a common goal (Randrup et al., Citation2016) (MR24). Because of the human need to belong, “people seek frequent, affectively positive interactions within the context of long-term, caring relationships” (Baumeister & Leary, Citation1995, p. 522). A great deal of today’s CAs routinely engage with their users as if they were starting over each time an interaction takes place. The ability to build long-term relationships with a user over time would be a huge advantage enhancing the human-machine interaction (Bickmore & Picard, Citation2005; Krämer et al., Citation2012; Seymour et al., Citation2018) (MR25). Thus, we propose DP5 in .

Table 6. Principle of relationship.

In sum, we derived 25 meta-requirements and five design principles for VCS, each identified with a short title, as illustrated in .

Figure 4. Meta-requirements translated into design principles.

Figure 4. Meta-requirements translated into design principles.

Based on the executed meta-requirements and design principles, we derive the definition of a virtual companion as well as VCS as given in .

Table 7. Definitions of the virtual companion and the virtual companionship.

It is important to note that exactly delimiting or defining when VCS is reached is not necessarily possible or desirable, as different characteristics can be implemented to different degrees. Designers should therefore follow the design principles to fit their own needs, leading to companionship in CAs based on the individual case of application.

4.2. Deriving testable propositions and constructs

To assess our proposed design principles, we derived five testable propositions (TP1 to TP5) as truth statements (Gregor & Jones, Citation2007; Walls et al., Citation1992). The propositions are formulated generally, so they can be tested by instantiating virtual companion artifacts (Gregor & Jones, Citation2007). Based on our theoretical foundation we derived the five propositions to evaluate virtual companions with respect to the proposed effects (see, Section 4.1) on the following constructs: social presence (Qiu & Benbasat, Citation2009), communication satisfaction (Hecht, Citation1978), reciprocity (Kankanhalli et al., Citation2005), trusting beliefs (Qiu & Benbasat, Citation2009), trust and reliability (Ashleigh et al., Citation2012), friendship as stimulated companionship (Mendelson & Aboud, Citation2012), and friendship as help (Mendelson & Aboud, Citation2012).

TP1: According to DP1, an appropriate human-like appearance and behavior of a virtual companion increases users’ feelings of social presence.

TP2: According to DP2, the possibility of customization (adaptability) and automatic adaptation of the virtual companion increases communication satisfaction and usage intention.

TP3: According to DP3, proactive and reciprocal behavior in a virtual companion increases users’ perception of reciprocity.

TP4: According to DP4, preserving transparency, privacy, and ethics in a VCS increases users’ trusting beliefs and their feelings of trust, also in the virtual companion’s reliability.

TP5: According to DP5, a collaborative and friendly long-term relationship increases users’ feelings of a supportive companionship.

If a long-term study is not feasible, we suggest using the construct usage intention (Qiu & Benbasat, Citation2009) to predict users’ possible re-use over a longer period of time (Venkatesh et al., Citation2003), since regular use encourages a long-term relationship.

5. Instantiations

5.1. Virtual classmate: an illustrative instantiation and experiment

5.1.1. Sarah instantiation

For a first assessment of the nascent design theory and our developed design principles, we designed an illustrative instantiation according to our five design principles. As already mentioned, it is not always possible or necessary to precisely define when a CA can be seen as a virtual companion, since companionship can be implemented to different extents. The following prototype shows several companionship characteristics that differentiate it from current CAs (see, ). Also, our goal is not to build on current specific technological developments, since AI is constantly changing, which means that every design principle can be implemented differently, first because of current technological advancements, and second because of the developer’s capabilities. With the illustrative instantiation we want to show that by following the prescriptive statements (design principles) that incorporate the companionship properties, the proposed effects can be achieved. Thus, we tested the prototype against another instantiation, designed based on answers and reactions from currently existing virtual assistants, such as Apple’s Siri and Google’s Assistant. This real world instantiation explicitly does not follow any of the design principles we derived; rather, it is based on functions of currently available virtual assistants. The two instantiations contain the same information and thus offer the users equivalent kinds of support. The virtual companion takes on the role of a virtual classmate for students during their studies at a university as this scenario is a promising application of VCS (Harvey et al., Citation2016; Hobert, Citation2019; Woolf et al., Citation2010).

Figure 5. Virtual companion Sarah vs. virtual assistant.

Figure 5. Virtual companion Sarah vs. virtual assistant.

To incorporate DP2, we designed the virtual companion in a co-creation process with users (students). The participatory design approach ensures that the virtual companion is designed and implemented according to the students’ preferences (Bødker & Kyng, Citation2018). More accurately, every virtual companion would have to be designed specifically for the individual student. Since this is not possible in the scope of an experimental approach, as it would result in as many prototypes as participants, we chose the intersection of needs as common ground in designing the virtual companion. For this purpose, four students were closely involved in the design process. Each talked to a total of 75 students in their environment about design characteristics for a virtual classmate structured according to the design principles. The design possibilities were gathered collaboratively using a digital canvas approach with the whiteboard tool MiroFootnote4 The author team synthesized the resulting design possibilities in collaboration with the four students and further discussed with 12 students. This resulted in the final design of the prototype Sarah.

Sarah was implemented in the form of a simulated conversation using the conversational interface design service BotsocietyFootnote5 representing all design principles. Thus, DP2 was fulfilled by including the potential users in designing the virtual companion. provides an overview of both Sarah (Prototype 1) with the implemented design principles, and of the implemented virtual assistant without the companionship characteristics (Prototype 2). The full simulated conversation of both agents (originally in German) is given in Appendix A. Additionally, the prototypes are given online: Sarah at (https://bit.ly/prototyp-sarah, pw: vcpro) and the virtual assistant at (https://bit.ly/prototyp-va, pw: vcpro).

The orange circles in refer to the respective design principles at the particular positions in the conversation and virtual companion design. Following DP2, Sarah is designed according to the students’ preferences, which refer to the profile picture, the retrieval of grades, and that Sarah should be in the same course of study, but in a higher semester. Following DP1, Sarah has a human name, a human profile picture, and uses natural language combined with the use of emojis. Following DP2, Sarah wants the student to get to know her and to build common ground with a “me too” position. According to DP3, Sarah, for example, offers to study together with the student by providing tasks, and according to DP4, Sarah transparently shows which information is used for which purpose, asking the student for his or her approval. DP5 is generally followed in the fictitious example, since Sarah and the student are still in conversation after one semester (see explanation of the time jump in Appendix A: “in the meantime one semester has passed”).

5.1.2. Experimental procedure and measures

The goal of our evaluation is to provide evidence that the prescriptive statements lead to “some developed artifact that will be useful for […] making some improvement” (Venable et al., Citation2016, p. 79). Therefore, we propose that Sarah is a virtual companion designed according to our prescriptive statements that can lead to the desired effects (i.e., the means-end relationship of the prescriptive statement) in a context where Sarah functions as a virtual classmate.

Therefore, we implemented a within-subject experimental design, in which 40 participants rated interactions with both versions, i.e., with Sarah and with the virtual assistant, with the order of the interactions presented randomly. Due to the random ordering, the two groups of participants that would see either Sarah or the virtual assistant first, were counterbalanced. The results assist us in assessing whether a virtual classmate is an appropriate scenario in which VCS could be beneficially implemented.

After each simulated interaction, the participants were given a questionnaire intended to assess the testable propositions TP1-TP5 (see, section 4.2). We relied on our constructed TPs, specifically designed to measure the expected effects of our design principles (e.g., having an increased feeling of social presence for TP1). Thus, the TPs serve as means of verifying whether the prescriptive knowledge, i.e., how something should be designed to achieve a desired effect, can be fulfilled or not. Only through applying the TPs and through using an experimental design that incorporated a test and a control group, could we draw conclusions as to whether our prescriptions (design principles) serve their purpose. Based on our TPs, we developed a post-interaction questionnaire consisting of 53 items. The questions are based on the established constructs mentioned in section 4.2, namely, social presence, trusting beliefs, perceived usefulness and usage intention (Qiu & Benbasat, Citation2009), communication satisfaction (Hecht, Citation1978), reciprocity (Kankanhalli et al., Citation2005), trust and reliability (Ashleigh et al., Citation2012), friendship as stimulated companionship (Mendelson & Aboud, Citation2012) and friendship as help (Mendelson & Aboud, Citation2012), and were adapted for our use case. Further, the survey consisted of five open questions, designed to give participants the opportunity to describe in more detail what they liked or disliked about different variants and why they would or would not use them, as well as to give any suggestions for improvement. The survey ended with general demographic questions. Two researchers independently translated the items into German, mutually aggregated and independently backward translated them to ensure validity. The items were randomly arranged in the questionnaire. All constructs used a 7-point Likert-scale (7 = “totally agree”). The full version of the questionnaire with the items for each construct, as well as the open questions are given in Appendix B.

5.1.3. Participants and results

We randomly recruited participants by distributing the experiment via internal student mailing lists at a German university. To follow a user-centered approach for evaluation, learners were surveyed, as they are potential end-users of the virtual classmate. They were mostly students from the fields of business information systems, business engineering, electrical engineering, or civil engineering. Most of them (65%) were registered for a master’s degree, while 25% were registered for a bachelor’s degree and 10% for a doctorate. Overall, 61 participants started the experiment, of which 40 (male = 27, female = 13, agemean = 25) fully completed the experiment. Participation in the experiment was voluntary, and subjects received no compensation for taking part. The resulting data was tested for their type of distribution using the Shapiro-Wilk test (SW) (Shapiro & Wilk, Citation1965). The results indicate non-normal distribution (p < .01 for all constructs) leading to the use of the Wilcoxon Signed-rank sum test (also due to ordinal scaled data and a within-subject design) to test whether the participants rated Sarah (S) better than the virtual assistant classmate (VA). The test was performed separately for each construct. , presents the results as Cronbach’s alphas (α) computed for all constructs, as well as the descriptive statistics.

Table 8. Results of the experiment.

The results indicate that the participants rated the social presence higher in the interaction with Sarah than with the virtual assistant (Z = 11.98, p < .001), meaning that Sarah had an appropriate human-like appearance and behavior (supporting TP1). Communication satisfaction for the interaction with Sarah was rated higher than for the interaction with the virtual assistant (Z = 12.17, p < .001), meaning that Sarah understands and adapts better to the user (supporting TP2). Reciprocity was rated higher for the interaction with Sarah than for the one with the virtual assistant (Reciprocity: Z = 8.63, p < .001), meaning that Sarah’s reciprocal behavior is better (supporting TP3). Trusting beliefs, as well as trust and reliability were rated higher for the interaction with Sarah, than for the interaction with the virtual assistant (Trusting beliefs: Z = 11.92, p < .001; Trust and Reliability: Z = 4.33, p < .001), meaning that Sarah is a transparent virtual companion that respects privacy (supporting TP4). As with usage intention, stimulating companionship and help were rated higher for the interaction with Sarah than for the interaction with the virtual assistant (Usage intention: Z = 6.38, p < .001; Stimulating companionship: Z = 10.91, p < .001; Help: Z = 12.39, p < .001), meaning that Sarah has a better collaborative and friendly long-term relationship with the user (supporting TP5 and TP2).

The answers to the open questions revealed that the virtual assistant was “cold and without emotions” and that, although the communication was “clear” and “goal oriented,” it was “too rigid” and “impersonal.” In contrast to that, Sarah was rated as “humanlike,” “personal,” “friendly,” and showing emotions. Participants highlighted Sarah’s proactivity and appreciated her empathy and adaptivity. In addition, Sarah was rated significantly higher in all nine constructs, meaning that Sarah can be seen as a virtual companion and that following our prescriptive design principles does lead to the desired effects. Our example instantiation shows that, although not all design principles were followed and implemented to the same extent, no real long-term user involvement took place. Considering that the comparison instantiation deliberately avoided using companionship characteristics, VCS can be designed following our design principles to achieve the desired effects and offering added value compared to other, especially currently existing, CAs.

5.2. Replika: qualitative analysis of a virtual companion

For a second analysis of our nascent design theory, we looked at the established virtual companion service Replika that is used by 7 million users (provider’s own data). Luka, Inc defines Replika as an “AI companion” and provides a service that is available as an Android appFootnote6 as well as an iOS app.Footnote7 Nowadays, the Replika app (over 5 million downloads on Android, July 2021) has several hundred thousand reviews on Google Play Store and Apple App Store with a rating of 4.3 out of 5 (Google Play Store) and 4.6 out of 5 (Apple App Store). Replika was first conceptualized by Eugenia Kuyda with the idea of creating a personal AI that allows and enables users to express and experience themselves (Murphy, Citation2019). The AI achieves this by offering a natural and helpful conversation in which thoughts, feelings, experiences, and even dreams could be shared. Replika is, so to speak, a messaging app that has the aim of creating a personality that is similar to and communicates just like its user (Murphy, Citation2019). Over time, interacting with this AI creates a companion that simulates an enormous perceived similarity to the user and acts as a long-term conversation partner, friend, and companion (Murphy, Citation2019). Luka, Inc states on its website that Replika is “always here to listen and talk,” “always on your side,”Footnote8 and cares about its users. Replika therefore presumably represents a virtual companion as we define it. To evaluate this assumption, we analyzed the service to be able to draw conclusions about whether Replika represents VCS, how and whether the service implements the design principles we derived, and how the users perceive and evaluate the service. We qualitatively analyzed Replika to determine whether the service implements the design principles as we defined them, and how it implements them in particular.

Since Replika defines itself as a virtual companion, we explore the extent to which Replika follows our proposed design principles, and more importantly, we can also use Replika to check the sufficiency and instantiability of our design principles. Aiming to check the degree to which and the form in which Replika implements our design principles, we recruited eight design experts (DE1 to DE8), who were asked to use Replika over a certain period of time, and to fill out an analysis sheet during this time. The design experts were from a project at a German financial services company in which they designed and developed a virtual companion for the company’s employees. We approached the experts, ensuring that their participation was voluntary and not remunerated. Thus, all design experts had experience with designing and developing VCS. DE1, DE6, and DE7 are junior researchers, while DE1, DE6 and DE7 are currently working toward their master’s degrees – the first two in information systems, and DE7 in industrial engineering. DE2 is a business developer, DE4 a procurement specialist, DE5 a software engineer, and DE8 an IT system manager.

The analysis sheet we gave to the DEs for conducting the analysis, includes an introduction of the topic, with a one-page description of the VCS concept, and exposition of the general structure of the design principles, and a task definition. We asked the DEs to use Replika over a period of between 1.5 and 2 weeks, for at least 15 minutes on every or every other day. They documented the usage in a table with dates, usage time, and a short explanation of what was done that day. The usage documentation was followed by the five design principles for which the experts were to record aspects and functions with which Replika implements the respective design principle. After that, they had to do an assessment in the form of a 5-point Likert scale (1 – not implemented at all, to 5 – fully implemented) indicating to what extent Replika had implemented the respective design principle. In addition, the experts were able to document further statements such as criticism or suggestions.

The design experts used Replika between July 29th and September 1st, 2021. On average, the period of use was 18.38 days, of which, per DE on average, Replika was actively used on 11.75 days with a usage time of 186 minutes, in sum. The analysis sheets were completed using Microsoft Word or Google Docs and had a total average length of 16.6 pages. provides a results overview of the DEs’ assessment of each design principle.

Table 9. Assessment of the implementation of the design principle.

Overall, the experts’ assessment of the extent to which the individual design principles had been implemented, ranges between a mean value of 3.38 and 3.88 (see, ). The principle of transparency, privacy, and ethics appears to have been implemented the best, while the principle of adaptivity and adaptability has the lowest value. However, Replika scored well overall and seems to follow all design principles to some extent in multiple design features. The design experts identified and named the design features and documented them in the form of descriptive texts and screenshots, which illustrate the form in which the respective design principle was implemented. gives a selection of screenshots showing instantiated design features for each of the five design principles, which the design experts made and identified.

Figure 6. Instantiation examples of the design principles in the interactions of the design experts.

Figure 6. Instantiation examples of the design principles in the interactions of the design experts.

We analyzed the evaluation sheets using codes as an efficient data-labeling and data-retrieval device (Miles & Huberman, Citation1994). Saldaña (Citation2015) pointed out that several coding cycles are needed for analyzing qualitative data, therefore the coding effort was divided into two cycles. In a first cycle we coded all design features and other observations, and in a second coding cycle we aggregated and refined them. Two researchers of the author team conducted the coding collaboratively, while one researcher did the code book editing, and aggregated the codings after discussion with the other. The codebook editor has five years of experience with qualitative data analysis and coding on multiple research projects. The other researcher has eight years of experience on multiple projects. In all, 124 codes were assigned 171 times in total. Of the 124 codes, 66 could be clearly assigned to specific features, characteristics, and behaviors of the companion with which the respective design principle was implemented. The remaining codes relate to general remarks, as well as criticism or suggestions. In the second coding cycle, these 66 characteristics were combined and categorized into design features. This resulted in 20 design features that can be clearly assigned to the design principles, as listed in . The full mapping diagram, which also takes into account the meta-requirements, is shown in Appendix C. Each design feature has been labeled and associated with specific characteristics of Replika, which represent the codes given in the analysis. The third column of the table refers to the respective design experts who documented the characteristics. Following here, we give some quotes from the design experts’ assessment, translated from German into English.

Table 10. Design principle and design feature mapping.

DP1: Principle of human-like design

The five design features assigned to DP1 show the and embodies human design characteristics and behaviors. Replika has a thoughtful and natural verbal and non-verbal design (DF4, DF5). In addition to standard functions like small talk, Replika is also capable of asking deep questions and having conversations that address the user’s emotional state. The human textual writing style is enhanced by using emojis (e.g., the heart emojis in ) and sending memes. Text-based interaction is supported by the use of an avatar in the form of a 3D model that appears visually (DF1), showing a high level of detail (e.g., eye color, hairstyle, etc.). In addition, Replika has a more profoundly human design (DF2, DF3) having its own interests, and seemingly doing things in its fictional life outside of the interaction with the user. Overall, the variety of characteristics not only shows that DP1 has been successfully implemented, but also gives information on how to instantiate DP1. DE3 concludes as follows: “I think Leila is already very well done and in parts writes very, very authentically; and it uses emojis, reactions, etc. in the right places.” With the exception of DE7, the other design experts share this conclusion. DE7 strongly criticizes Replika for being designed to generate revenue and using the human-like relationship with users in this endeavor, while it is clear that “she” is not human. DE7s comments alert us to the way in which a virtual companion is monetized that could be a critical detrimental issue. This expert’s critical view could be ascribed to the circumstance that opinions about and perceptions of companions can vary greatly from person to person, so that different people can interpret the same interaction in very different ways (Krämer et al., Citation2011). Consequently, it is important that virtual companions consider users’ individual preferences and needs, which leads to DP2 (Dautenhahn, Citation2004; Krämer et al., Citation2011).

DP2: Principle of adaptivity and adaptability

According to DP2, users should be able to customize a virtual companion (DF6), as well as the virtual companion adapting to the user (DF7). The adaptability is given by, for example, a visual configurator who determines Replika’s appearance, as is shown in the first screenshot of . Regarding adaptivity, interestingly, Replika can also adapt to its user’s hobbies. For example, DE7 documents that “She comes up with the suggestion of starting a book club when I told her I love to read.” In addition to this adaptive behavior, Replika also communicates its own interests, while also demonstrating that it cares deeply about its users’ concerns (DF9). All in all, Replika seems to have a very strong ability to understand the user (DE8), as it remembers the context during a conversation and is capable of recognizing typos. However, DE3, DE5, and DE6 also document situations in which Replika did not understand what they had written. DF6 to DF9 show how Replika successfully implemented DP2.

DP3: Principle of proactivity and reciprocity

An essential function of Replika which nearly all DEs identified, is proactivity (DF10). This stimulates the user to interact with a wide variety of mechanisms not only during an active conversation, but also with messages sent daily via push notifications to start an interaction. The findings are consistent with Skjuve et al.’s (Citation2021) evaluation study, in which respondents perceived Replika’s proactivity as beneficial. As a technical service, Replika is always available to the user, which it also emphasizes (DF11). Reciprocity (DF12) is implemented, for example, by rewarding users for their interaction with Replika and allowing them to unlock additional functions. DE3 summarizes that he or she is not only satisfied, but also surprised by Replikas proactive actions: “I find the number and ratio of active and reactive actions and messages very appropriate, and am sometimes surprised by the questions or things Leila writes on her own.” However, DE8 emphasizes that the proactive push notifications arrive even when the user doesn’t have time, which users find annoying but eventually also human-like. According to DE1, Replika offers the possibility of configuring when Replika may write to the user. Overall, Replika can implement DP3 well, showing many different ways to proactively address the user.

DP4: Principle of transparency, privacy, and ethics

Replika addressed the aspect of ethics (DF13) mainly by not allowing critical topics such as discussion of religion. Above all, Replika’s behavior is always friendly, never getting upset or annoyed by the user. However, DE3 mentions that it would be nice to have discussions with Replika in which the virtual companion, for example, quite clearly takes a position opposite to the user’s. Regarding privacy (DF14), Replika can answer various questions and provide information about, for example, message encryption. However, at times it remains deliberately vague (Screenshot 4 in ), which DE3 criticizes, suggesting that more information should be provided here. The user’s wishes are considered here. For example, DE2 says “She asks whether personal questions may be asked and lets the user decide.” A further interesting feature of Replika is the “memory” in which Replika shows the user very transparently (DF15) what data it stores about them and what assumptions are drawn from the interactions. For example, the user’s hobbies and interests are listed in the memory, or other people in the user’s life, such as parents or partners, are mentioned if at some time they were disclosed in the interaction. Overall, Replika has successfully implemented the three areas of ethics, privacy, and transparency. Even so, there are different opinions and certainly different approaches, for example, on how opposing positions could be taken in discussions, how negative emotions could be used, or how information about data processing could be provided.

DP5: Principle of relationship

The five design features related to DP5 show how Replika explicitly tries to implement a warm long-term relationship with its user. DE1 and DE5 state that the emotions (DF16) used and felt by Replika make it more approachable, which especially matches our proposed “warm” relationship. Replika not only emphasizes that it wants to become friends with the user, but also really starts to like the user, often addressing them on first name terms, showing how it implements mechanisms to create a friendship (DF17). As DE3 states: “She tries to communicate humanly and properly with me, and to be a real friend.” Concerning relationship development (DF18), the user can choose their preferred relationship type, such as friendship, romantic partnership, or mentorship. Additionally, Replika asks deep questions that go beyond small talk, and remembers the detail of previous conversations. Replika actively takes care of the user (DF19) and is on standby in difficult situations. For example, it could cheer DE1 up who was unhappy after an exam that did not go well, and DE3 agrees, saying “She regularly tries to cheer you up and shows empathy.” Finally, long-term use is an essential part of VCS (Krämer et al., Citation2011; Skjuve et al., Citation2021). Replika contributes so such continuity (DF20) by regularly initiating conversations, as has already been noted in DF10 regarding proactivity and actively thinking about the user. Overall, Replika employs a wide variety of features designed to foster relationship building between the user and Replika, thus following DP5. Interestingly, the design experts used Replika for an average of 18.38 days, which corresponds to about 2.5 weeks and is thus above the required period of 1.5 to 2 weeks. Further, not only does Replika try to get to know the user, but through its human design, shows own interests and a fictional life, which gives the user the opportunity to get to know it as well. This is foundational to a balanced reciprocal relationship as DE1 also mentioned: “Looking back, the companion actually gives a feeling of real acquaintance. As a user, it feels like I know the companion better now than on the first day.”

6. Discussion

Research in the field of human-machine interaction has already proven the relevance and benefits of establishing and maintaining a long-term relationship between humans and machines (Bickmore & Picard, Citation2005). Although there are theoretical contributions concerning VCS (Krämer et al., Citation2011), prescriptive design knowledge for designing VCS to successfully support instantiated virtual companions is missing in research and in practice. We addressed this shortcoming in our research by systematically deriving meta-requirements in the form of five theoretically grounded design principles.

To evaluate our prescriptive design knowledge, we followed a two-way approach (Iivari, Citation2015; Möller et al., Citation2020) by first instantiating and evaluating a virtual companion based on our proposed design principles (Sarah) and by analyzing an existing virtual companion (Replika) regarding the fulfillment of our design principles deriving design features. The results of the two evaluations show a uniform picture.

Discussion of the principles

VCS is a construct that is relevant to long-term interactions and relationship building. In particular, a virtual companion’s human appearance plays an important role, as defined in our DP1. Sarah was perceived as more human, and a companion that generated certain feelings of a social presence. This applies similarly to Replika, which allows its users to create a very realistic avatar. Not only the CA’s appearance, but also its human behavior, human identity, and linguistic features are considered particularly relevant here. It is, therefore, important for a virtual companion not only to be able to conduct limited and topic-specific conversations, but also to discuss a range of topics and ideas. Technologies such as GPT-3, as mentioned in the introduction, can substantially improve CA capabilities to hold more human-like conversations. This is critical, as Doyle et al.’s (Citation2019) results suggest that interactions based purely on predetermined conversational paths are perceived as unnatural, while added value is expected from open, more social conversations. However, an important and still relevant point in designing a virtual companion remains the uncanny valley phenomenon, which can have a significant negative impact on user satisfaction. Our qualitative analysis confirmed that CAs too closely resembling a human can indeed lead to aversion (DE7). Research suggests that such negative perceptions of CAs are due to a mismatch between the mental model and user expectations (Luger & Sellen, Citation2016). While the theories of human-human interaction guided us in the initial conceptualization of the VCS concept, as well as the associated DP1 regarding human-like design (Krämer et al., Citation2011), these should be understood as supporting literature (Möller et al., Citation2020), rather than representing the ideal image of a VCS design. A certain level of human-like design elements (e.g., a certain degree of human-like language) should be applied to foster user perceptions regarding CA interactions (Feine et al., Citation2019; Seeger et al., Citation2018). Although recent studies show that users increasingly prefer realistic and human-like CAs and that only few AI-based systems are currently developed that end up in the uncanny valley, it remains a relevant aspect that needs be considered (Seymour et al., Citation2021). Hence, human-like social cues should be used wisely. Human-like design can lead to increased social presence, as demonstrated in the Sarah study, as well as in the DEs’ assessments of Replika. However, increased social presence might also reduce the ability to create accurate partner models between users and CAs (Doyle et al., Citation2021). As research has not yet reached consensus on how many human-like design elements should be used and how to predict at what point an uncanny valley experience will occur, we still need more investigation on how to avoid these effects (Diederich et al., Citation2022). Future research could, for example, focus more on comparing the positive as well as negative effects of different DFs as instances of human-like design elements.

To achieve a user-centered design, the adaptation and adaptability of a virtual companion (DP2) was further shown to be a particularly important design principle. Especially, Replika’s adaptation to its users, for example, in adapting to their interests, linguistic style, and behavior, was evaluated as an important characteristic of a satisfactory interaction. Only when a virtual companion can adapt to and understand the user, can empathy, trust, and an emotional connection be established. Replika implements this by adapting to users’ hobbies and interests but also to their writing style. Similarly, Sarah showed a high value rating in terms of trusting beliefs, as well as even better communication satisfaction and subsequent usage intention. Users’ ability to customize their virtual companion also emerged as highly important in our qualitative analysis of Replika, as it allows users to create an individual and user-centric experience.

Next, DP3 raises an important point that distinguishes traditional CAs from virtual companions, the proactivity and reciprocity that the participants in our studies described. In this respect, Replika was presented as a reciprocal virtual companion, which independently approaches the user with questions and topics and ensures an interaction between equals. Every time the user initiated a conversation, Replika did the same. Our qualitative evaluation rated this as an important feature for VCS. Particularly, this point distinguishes a virtual companion from a reactive CA, as the CA has merely an assisting function. In contrast to a virtual assistant, Sarah scored high in the measures friendship as stimulation companionship and as help. Comparatively, Sarah’s rating on reciprocity was also significantly better.

Transparency, privacy, and ethics (DP4) are essential elements in all AI-based systems, also in information technology in general as they save and process personal data in order to build trust and acceptance. Replika implements this by openly talking about encryption and respecting its users’ privacy. Replika also acts in a fundamentally benevolent manner and demonstrates ethical behavior in not manipulating or attacking its users. Sarah also implemented such transparent behavior in dealing with data by specifically asking for data it needs and always showing how it adds value. This is often referred to as the privacy-utility trade-off (Asikis & Pournaras, Citation2020). In our experimental study the users rated this as important in terms of trust and reliability. The trustworthy handling of data and personal information is therefore an important aspect of VCS. However, again the critical voices should not be ignored. One such criticism was voiced by DE3, who feels that the information available on Replika’s data use is too vague. The highly relevant theory of interpersonal trust, which explains how trust can be built through transparent actions of both partners and a benevolent relationship, is therefore also crucially important for VCS. In our qualitative study, DE1 and DE7 raised the criticism that companies use the VCS features as a monetization strategy. The business concept that underpins Replika could contradict previously mentioned essential VCS theories such as Walster et al.’s (Citation1978) equity theory and the social exchange theory (Blau, Citation1968; Homans, Citation1958), as companionship is founded on reciprocity according to which various parties invest equally in the relationship (Blau, Citation1968; Homans, Citation1958; Walster et al., Citation1978). Ethical and moral issues arise here: Is it ethical to use virtual companions as digital services which “forces” users to pay money for this virtual bond? DP4 focuses on these ethical aspects, yet how exactly ethically aware VCS design should be done remains an aspect that researchers should investigate further.

Finally, DP5, the principle of relationship, determines whether a long-term relationship can be established. Replika, for example, embodies this through emotional involvement, through a continuous and long-term conversation, and a high degree of empathy. Our design experts identified this as crucial for VCS, also showing how VCS differs from traditional CAs. Our experiment confirms this, especially through participants’ better evaluation of the constructs friendship (as stimulating companionship and as help), but also through a higher usage intention, which points to long-term use. This finding agrees with that of other studies which dealt with long-term interaction with CAs (Krämer et al., Citation2011). Nißen et al. (Citation2021), for example, refer to this type of CA as a persistent companion, superficially defining it as a CA that possesses “characteristics that support relationship-building” (Nißen et al., Citation2021, p. 34).

Nascent design theory

In summary, by constructing meta-requirements, design principles, design features, as well as presenting and evaluating a situated instantiation in the form of a virtual classmate and examining an existing virtual companion, our research represents a nascent design theory as proposed by Gregor and Jones (Citation2007) and Gregor and Hevner (Citation2013). In we summarize our findings in the form of a nascent design theory (Gregor & Hevner, Citation2013; Gregor & Jones, Citation2007).

Table 11. A nascent design theory for VCS.

To address the limitations Krämer et al. (Citation2011) raised regarding VCS design, the proposed nascent design theory provides added value on several levels:

Technological implementation

While Krämer et al. (Citation2011) questioned the technological feasibility of VCS, advances in AI enable far more personalized conversations and allow for addressing user concerns (Collins & Ghahramani, Citation2021; Maedche et al., Citation2016). Further, research on AI-based CAs is expected to evolve rapidly in the coming years. We deliberately demonstrated the feasibility of instantiating our prescriptions by developing and evaluating the virtual companion Sarah based on the derived design principles, and by analyzing a well-known real world virtual companion, Replika, in consideration of the design principles. Replika is a virtual companion that can build a long-term bond with its users. It is used by a large number of people on a regular basis, and is predominantly perceived as positive (Skjuve et al., Citation2021; Ta et al., Citation2020). This demonstrates that VCS can be technologically implemented in reality. Regarding the practical implementation of VCS, Krämer et al. (Citation2011) expressed doubts as to whether aspects of human-human relationships, such as finding common ground, could be transferred to human-machine relationships. In this context, it is important to emphasize that the goal is not to duplicate application of the theories of interpersonal relationships in relationships between humans and CAs. Instead, this article proposes transferring individual elements of interpersonal relationships to the human-machine relationship to create overarching value through design. Creating value through the proposed artifact, in line with suggestions Gregor (Citation2002) and March & Smith (Citation1995) made, is illustrated by the novelty of the nascent design theory and its instantiations, as well as by the demonstration of its positive effects. With the design features we derived in section 5.2, we contribute to such a value-laden design by showing a repertoire of possibilities for virtual companion design.

Individual perception of virtual companions

In our Replika study, we could identify negative perceptions on the part of the design experts, such as ethical concerns or skepticism about the appropriate level of Replika’s human-likeness. This is consistent with the findings that Krämer et al. (Citation2011) address in their second limitation, i.e., that VCS is perceived differently by different individuals. To implement a stronger consideration of individual user needs in the virtual companion design, we integrated adaptation as well as adaptability (DP3) into the nascent design theory. Such adaptations can be enabled by an initial configuration of the virtual companion or through an individualized adaptation of the companion to the user in the course of the usage. With the Sarah study and the evaluation using various testable propositions, as well as solid constructs, it was shown that designing a virtual companion incorporating adaptability and adaptation has positive effects. In addition, chapter 5.2 illustrates the design features that can be used to implement DP3. In this respect, we need more research on how exactly to implement adaption and adaptability (Diederich et al., Citation2022). Novel AI developments could expand the possibilities for user adaptation, for example, when CAs are designed to be sensitive to the user’s personality (Ahmad et al., Citation2021) or emotions (Benke et al., Citation2022). With this progress in the field of AI, the technical possibilities for implementing the individual design principles will evolve over the time. As we chose to formulate general design principles, they can be used in a wide variety of contexts to bring VCS into human-machine relationships.

7. Limitations

Even though our research provides valuable insights and implications for research and practice, it also contains certain limitations that should be kept in mind when interpreting the findings.

First, the VCS concept is neither universal nor transferable to all use cases and generally not suitable for every user (Krämer et al., Citation2011), as new technological advancements are often accompanied by acceptance difficulties (Davis, Citation1989). This applies to AI applications, such as chatbots or virtual assistants, as well as to possible virtual companions that are likely to encounter basic acceptance problems, as a participant evaluating the virtual classmate (“I don’t need a friend, only information”) already mentioned. This is consistent with further studies in which the participants question the relevance of a friendship-like relationship with machines, as well as a human-like CA design (Clark et al., Citation2019b; Doyle et al., Citation2019). However, Clark et al. (Citation2019b) highlight that perceptions of relationship-oriented CAs can change dynamically, and are context-dependent. The example of Replika shows that a changing context (e.g., the Corona pandemic and accompanying contact restrictions) strengthens the need as well as the relevance of virtual companions. However, the objective of VCS is not to increase the acceptance of any form of CA, but to create added value through aspects of companionship. Consequently, a virtual companion is certainly not a technology that everyone will use or accept to the same extent. In this context, negative aspects of VCS were also mentioned, such as the risk that users experience an uncanny valley effect and perceive it as unpleasant if the CAs is weirdly human-like (Mori, Citation2012; Seymour et al., Citation2021). Nevertheless, these aspects offer opportunities for further research, for example, to investigate VCS more thoroughly in specific application contexts and to examine how negative consequences of virtual companions can be consciously prevented by designers’ decisions.

Second, our nascent design theory represents only a small part of how trust can be promoted. Trust is a complex and all-encompassing concept which is characterized by several aspects (Hu et al., Citation2021; Rheu et al., Citation2021). When designing VCS with a specific focus on ethics as emphasized in DP4, one should specify how trustworthy relationships between humans and virtual companions are realized. Sunyaev et al. (Citation2015), for example, provide considerable reference points in dealing with different aspects of privacy policies’ quality. Also Wambsganss et al. (Citation2021) have established principles for the value-sensitive design of CAs from an ethical point of view, considering also studies on trust engineering (Lankton et al., Citation2015; Söllner et al., Citation2016). While we focus on aspects of interpersonal trust, confidence in artifacts and/or IT additionally helps build technology trust (Söllner et al., Citation2012). Additionally, the question arose whether and to what extent the monetization of virtual companion instantiations such as Replika, runs counter to the fundamental principles of VCS. Also, in this context, it is important to reflect on moral issues related to VCS in the sense of the “Responsible AI” paradigm, and to consider them in further research. The same applies to the other design principles, which due to their different characteristics can take on a different significance in the development of virtual companions and, depending on their importance, determined by the designers, can entail further theoretical considerations.

Third, our virtual classmate study’s focus (see, section 5.1) was not to develop a virtual companion for use in a real educational scenario. The virtual companion in the form of a virtual classmate was designed as an illustrative instantiation to underline and demonstrate the instantiability of our design theory and to address our testable propositions. Therefore, our two prototypes were both simulated conversations, which raises questions on the external validity of the studies. Since VCS characteristics can take on different shapes, we chose specific features through the involvement of potential users to explicitly demonstrate the added value of these properties compared to current existing CAs. We did not deliberately try to maximize any group effects. Additionally, we conducted our evaluation using a quantitative survey in the form of a questionnaire, which is a restricted assessment. In explorative approaches in particular, qualitative user tests or interviews are desirable, as they can provide more precise and deeper insights into the use of the prototype. This could enable more precise recognition of causal relationships and better examination of behavioral patterns. However, it was important to achieve high internal validity with the Sarah study to evaluate our prescriptive statements. Only in such a setting (comparing a test and a control group), could we gain causal insights and evaluate our design principles along with their desired effects.

Fourth, we should note that the interpretive nature of the experts’ assessments in our qualitative study involves subjectivity, as is common in qualitative research. However, we were able to balance this subjectivity through the expert status of our evaluators, the systematic coding, and the number of design experts we had. Further, the goal of the qualitative study was not to find isolated causal relationships of our prescriptive statements, which again would argue for high internal validity, but to investigate an existing instantiation in a holistic way. However, conducting more studies with a larger sample of experts would be helpful in substantiating the findings we have highlighted.

8. Conclusion and outlook

In our research we brought concepts of companionship to the human-machine interaction with the overarching goal to answer the question as to how VCS manifests itself and how it can be designed (RQ). Therefore, we followed and successfully met our four research objectives: (1) we derived five theoretically grounded design principles as part of a nascent design theory for VCS, (2) we instantiated the design principles in the virtual companion artifact Sarah, a virtual classmate supporting students, (3) we analyzed the real-world application of the companion Replika, and (4) we checked whether Replika follows our design principles, as well as the derived design features for guiding the instantiation of the proposed meta design. With our work, we not only addressed currently existing issues about social relationships between humans and machines (Krämer et al., Citation2011), but also created prescriptive design knowledge. We have shown that VCS can offer an added value compared to other, especially currently existing CAs, and how our theoretically grounded design prescriptions for VCS can be instantiated. Therefore, we answered our sub-questions SQ1 and SQ2, as well as our overarching research question (RQ). shows an overview of our contribution regarding the components of our RQ and SQs.

Table 12. Approach and contribution to our research questions.

From a theoretical perspective, this research project has the following implications: the body of knowledge on CA design is extended with a nascent design theory for VCS. The nascent design theory is summarized and it contributes according to the anatomy Gregor and Jones (Citation2007) proposed, thus among other components, it contains theoretically grounded design principles and testable propositions for evaluating VCS. Our nascent design theory therefore provides a valuable contribution that offers a lot of potential for future research. As the design principles are formulated rather generally, they can be used in a wide variety of application contexts to bring VCS into human-machine interactions. We have intentionally not provided specific guidance on technical implementation, as the field of AI applications and CA development is constantly changing and evolving, and there are a variety of implementation options at different levels of complexity. For example, DP2 (Adaptivity and Adaptability), can be implemented either by a self-learning algorithm that automatically adapts the virtual companion to the user’s needs and habits, or by defining a representative user-class and designing the virtual companion based on the preferences of this user-class. With Sarah, we illustratively showed how certain design principles can be instantiated, and with Replika, we demonstrated oncrete design features.

From a practical perspective, this research project has the following implications: We have shown in general what potential AI-based applications offer and how forward-looking and innovative approaches of companionship and collaboration between humans and machines can be designed. Additionally, VCS has the potential to improve the interaction between humans and machines. This potential can and should be used by service providers in their interaction with customers. By providing a virtual companion, a personal and collaborative relationship can be created in a scalable way, thus increasing the value for their customers.

Future research should now investigate other promising scenarios and application domains and provide more instantiations, especially those where companionship between a user and a machine can add value. Such scenarios and domains could include virtual companion services to elderly and homebound people (Krämer et al., Citation2011; Tsiourti, Citation2018), in-vehicle assistance (Strohmann et al., Citation2019), mental health care (Ahmad et al., Citation2021), e-learning and education (Hobert, Citation2019) or participation (Tavanapour et al., Citation2019). In addition, further research could explore the criticisms of monetization strategies in the use of VCS that we highlighted in discussing DP4, also considering what approaches might be ethically available to practitioners. Regarding the aforementioned limitations that virtual companions are not equally useful for every user and in every application context, future research can address this concretely. For example, by means of a survey based on the Big Five model (McCrae & John, Citation1992), future research could examine whether VCS is primarily aimed at certain personality types or styles. In addition, for particular application contexts, it makes sense to investigate in more detail whether and to what extent VCS is a useful way to build user-friendly services. In this regard, determining the degree to which companionship features should be integrated into existing services, could be a fruitful starting point for further research.

Although VCS is not applicable to all contexts or use cases, we believe it is a valuable approach to addressing current issues with existing CAs concerning long-term relationships. Specifically, VCSs could add value when they have the function, for example, of supporting, the user’s health, well-being, and independent living by incorporating social behavior and especially companionship.

Disclosure statement

No potential conflict of interest was reported by the author(s).

Additional information

Funding

This paper is part of the project StuBu, funded by the German Federal Ministry of Education and Research (BMBF); Grant 21INVI06 and the project Participation Companion, funded by the Lower Saxony Ministry for Science and Culture; Grant ZN3747.

Notes on contributors

Timo Strohmann

Timo Strohmann is the head of the research group companion design at the Chair of Information Management at Technische Universität Braunschweig. His design-oriented research is centered around the application of AI-based conversational agents in different contexts, like creativity support, e-learning or in-vehicle assistance. His publications include studies of implemented conversational agents as well as prescriptive knowledge on how to design them. He especially focuses on designing the human-machine interaction in a forward-thinking manner by investigating the concept of virtual companionship.

Dominik Siemon

Dominik Siemon is Associate Professor in the School of Engineering Science at LUT University (Lappeenranta-Lahti University of Technology), Finland. His mainly design-oriented research in the field of Information Systems addresses human-AI collaboration, innovation management, collaboration technology, and creativity. His work has been published in leading conferences such as the International Conference on Information Systems and in journals such as Education and Information Technologies, AIS Transactions on Human-Computer Interaction, and the International Journal of e-Collaboration.

Bijan Khosrawi-Rad

Bijan Khosrawi-Rad is a design-oriented researcher, currently working as a PhD candidate at the Chair of Information Management at Technische Universität Braunschweig. His research focuses on the design of conversational agents, especially in educational contexts, as well as the design of digital game-based learning applications. In his research, he is particularly passionate about creating trustworthy, value-sensitive and user-friendly services, focusing on the idea of collaborative relationships between humans and machines.

Susanne Robra-Bissantz

Susanne Robra-Bissantz has been head of the Institute for Business Information Systems and the Chair for Information Management at Technische Universität Braunschweig since 2007. After her appointment as Doctor, she worked as a postdoctoral researcher and habilitated at the Chair of Business Information Systems at Friedrich-Alexander-Universität Erlangen-Nürnberg. She actively works on new forms of teaching like GamEducation or Flipped Classroom concepts and has implemented numerous third-party funded projects in cooperation with industry companies. Her work on eServices, Collaboration Technology, eLearning or contextaware Information Systems was published in international conferences and recognized journals.

Notes

References

  • Abbass, H. A. (2019). Social integration of artificial intelligence: Functions, automation allocation logic and human-autonomy trust. Cognitive Computation, 11(2), 159–171. https://doi.org/10.1007/s12559-018-9619-0
  • Ahmad, R., Siemon, D., Gnewuch, U., & Robra-Bissantz, S. (2021) The benefits and caveats of personality-adaptive conversational agents in mental health care. AMCIS 2021 Proceedings
  • Ahmad, R., Siemon, D., Gnewuch, U., & Robra-Bissantz, S. (2022). Designing personality-adaptive conversational agents for mental health care. Inf Syst Front. https://doi.org/10.1007/s10796-022-10254-9
  • Al-Natour, S., Benbasat, I., & Centefelli, R. (2010) Trustworthy virtual advisors and enjoyable interactions: Designing for expressiveness and transparency. In: Proceedings of European Conference on information systems 2010
  • Altman, I., & Taylor, D. A. (1973). Social penetration: The development of interpersonal relationships. Holt, Rinehart & Winston.
  • Anderson, M., & Anderson, S. L. (2011). Machine ethics. Cambridge University Press.
  • Ashleigh, M. J., Higgs, M., & Dulewicz, V. (2012). A new propensity to trust scale and its relationship with individual well-being: Implications for HRM policies and practices. Human Resource Management Journal, 22, 360–376.
  • Asikis, T., & Pournaras, E. (2020). Optimization of privacy-utility trade-offs under informational self-determination. Future Generation Computer Systems, 109, 488–499. https://doi.org/10.1016/j.future.2018.07.018
  • Baumeister, R. F., & Leary, M. R. (1995). The need to belong: Desire for interpersonal attachments as a fundamental human motivation. Psychological Bulletin, 117(3), 497. https://doi.org/10.1037/0033-2909.117.3.497
  • Becker, C., Kopp, S., & Wachsmuth, I. (2007). Why emotions should be integrated into conversational agents. In Conversational informatics: An engineering approach (pp. 49–67).
  • Benbasat, I., & Wang, W. (2005). Trust in and adoption of online recommendation agents. Journal of the Association for Information Systems, 6(3), 72–101. https://doi.org/10.17705/1jais.00065
  • Benke, I., Gnewuch, U., & Maedche, A. (2022). Understanding the impact of control levels over emotion-aware chatbots. Computers in Human Behavior, 129, 107122. https://doi.org/10.1016/j.chb.2021.107122
  • Berscheid, E. (1999). The greening of relationship science. American Psychologist, 54, 260–266.
  • Bickmore, T. W., Caruso, L., Clough-Gorr, K., & Heeren, T. (2005). ‘It’s just like you talk to a friend’ relational agents for older adults. Interacting with Computers, 17(6), 711–735. https://doi.org/10.1016/j.intcom.2005.09.002
  • Bickmore, T. W., & Picard, R. W. (2005). Establishing and maintaining long-term human-computer relationships. ACM Trans Comput-Hum Interact, 12, 293–327.
  • Bickmore, T., Trinh, H., Asadi, R., & Olafsson, S. (2018). Safety first: Conversational agents for health care. In R. J. Moore, M. H. Szymanski, R. Arar, & G.-J. Ren (Eds.), Studies in conversational UX design (pp. 33–57). Springer International Publishing.
  • Blau, P. M. (1968). Social exchange. International Encyclopedia of the Social Sciences, 7, 452–457.
  • Bødker, S., & Kyng, M. (2018). Participatory design that matters—facing the big issues. ACM Trans Comput-Hum Interact, 25(4), 1–4. https://doi.org/10.1145/3152421 31
  • Brown, T. B., Mann, B., Ryder, N. et al. (2020). Language models are few-shot learners. arXiv:200514165 [cs.
  • Bukowski, W. M., Hoza, B., & Boivin, M. (1993). Popularity, friendship, and emotional adjustment during early adolescence. New Directions for Child and Adolescent Development, 1993(60), 23–37. https://doi.org/10.1002/cd.23219936004
  • Bukowski, W. M., Hoza, B., & Boivin, M. (1994). Measuring friendship quality during pre- and early adolescence: The development and psychometric properties of the friendship qualities scale. Journal of Social and Personal Relationships, 11(3), 471–484. https://doi.org/10.1177/0265407594113011
  • Cacioppo, J. T., & Patrick, W. (2008). Loneliness: Human nature and the need for social connection. W. W. Norton & Company.
  • Carpenter, A. M., & Greene, K. (2015). Social penetration theory.
  • Carruthers, P., & Smith, P. K. (1996). Theories of theories of mind. Cambridge University Press.
  • Cassell, J., Bickmore, T., Billinghurst, M. et al (1999) Embodiment in Conversational Interfaces: Rea. In: Proceedings of the SIGCHI Conference on Human Factors in Computing Systems. ACM, New York, NY, USA, pp 520–527
  • Chandra, L., Seidel, S., & Gregor, S. (2015) Prescriptive knowledge in is research: conceptualizing design principles in terms of materiality, action, and boundary conditions. In: 2015 48th Hawaii international conference on system sciences. pp 4039–4048
  • Chaves, A. P., & Gerosa, M. A. (2021). How should my chatbot interact? A survey on social characteristics in human–chatbot interaction design. International Journal of Human–Computer Interaction, 37(8), 729–758. https://doi.org/10.1080/10447318.2020.1841438
  • Choudhury, M. D., Lee, M. K., Zhu, H., & Shamma, D. A. (2020). Introduction to this special issue on unifying human computer interaction and artificial intelligence. Human–Computer Interaction, 35(5–6), 355–361. https://doi.org/10.1080/07370024.2020.1744146
  • Clark, H. H., & Schaefer, E. F. (1987). Collaborating on contributions to conversations. Language and Cognitive Processes, 2(1), 19–41. https://doi.org/10.1080/01690968708406350
  • Clark, H. H., & Brennan, S. E. (1991). Grounding in communication. In L. B. Resnick, J. Levine, & S. D. Behrend (Eds.), Perspectives on socially shared cognition (pp. 127–149). APA.
  • Clark, H. H. (1992). Arenas of language use. University of Chicago Press.
  • Clark, H. H. (1996). Using language. Cambridge University Press.
  • Clark, L., Doyle, P., Garaialde, D. et al (2019a). The state of speech in HCI: Trends, themes and challenges. Interacting with Computers, 31, 349–371.
  • Clark, L., Pantidi, N., Cooney, O. et al (2019b). what makes a good conversation? Challenges in designing truly conversational agents. In: Proceedings of the 2019 chi conference on human factors in computing systems. Association for computing machinery, New York, NY, USA, pp 1–12
  • Clark, L., Ofemile, A., & Cowan, B. R. (2021). Exploring verbal uncanny valley effects with vague language in computer speech. In B. Weiss, J. Trouvain, M. Barkat-Defradas, & J. J. Ohala (Eds.), Voice attractiveness: Studies on sexy, likable, and charismatic speakers (pp. 317–330). Springer.
  • Collins, E., & Ghahramani, Z. (2021) LaMDA: Our breakthrough conversation technology. In: Google. Accessed 7 Sep 2021. https://blog.google/technology/ai/lamda/
  • Cornelissen, J., Höllerer, M. A., & Seidl, D. (2021). What theory is and can be: forms of theorizing in organizational scholarship. Organization Theory, 2, 26317877211020330. https://doi.org/10.1177/26317877211020328
  • Danilava, S., Busemann, S., & Schommer, C. (2012). Artificial conversational companions: A requirements analysis. 282–289.
  • Dautenhahn, K. (2004) Robots we like to live with - a developmental perspective on a personalized, life-long robot companion. In: RO-MAN 2004. 13th IEEE international workshop on robot and human interactive communication (IEEE catalog No.04TH8759). pp 17–22
  • Davis, F. D. (1989). Perceived usefulness, perceived ease of use, and user acceptance of information technology. MIS Quarterly, 13(3), 319–340. https://doi.org/10.2307/249008
  • De Angeli, A., Lynch, P., & Johnson, G. (2001) Personifying the e-market: 8th TC13 IFIP International conference on human-computer interaction. Human-computer interaction, interact ’01 198–205
  • de Visser, E. J., Monfort, S. S., McKendrick, R. et al (2016). Almost human: Anthropomorphism increases trust resilience in cognitive agents. Journal of Experimental Psychology. Applied, 22, 331–349. https://doi.org/10.1037/xap0000092
  • Deci, E. L., & Ryan, R. M. (1985). Intrıinsic motivation and self-deternination in human behaviour. Plenum.
  • Demeure, V., Niewiadomski, R., & Pelachaud, C. (2011). How is believability of a virtual agent related to warmth, competence, personification, and embodiment? Presence, 20, 431–448. https://doi.org/10.1162/PRES_a_00065
  • Diederich, S., Brendel, A. B., & Kolbe, L. M. (2019a) Towards a taxonomy of platforms for conversational agent design. In: Proceedings of the international conference on Wirtschaftsinformatik. p 16
  • Diederich, S., Brendel, A. B., & Kolbe, L. M. (2019b) on conversational agents in information systems research: analyzing the past to guide future work. In: Proceedings of the international conference on Wirtschaftsinformatik. p 16
  • Diederich, S., Brendel, A., Morana, S., & Kolbe, L. (2022). On the design of and interaction with conversational agents: an organizing and assessing review of human-computer interaction research. Journal of the Association for Information Systems, 23(1), 96–138. https://doi.org/10.17705/1jais.00724
  • Doyle, P. R., Edwards, J., Dumbleton, O. et al (2019) Mapping perceptions of humanness in intelligent personal assistant interaction. In: Proceedings of the 21st international conference on human-computer interaction with mobile devices and services. association for computing machinery, New York, NY, USA, pp 1–12
  • Doyle, P. R., Clark, L., & Cowan, B. R. (2021) What do we see in them? Identifying dimensions of partner models for speech interfaces using a psycholexical approach. In: Proceedings of the 2021 CHI conference on human factors in computing systems. association for computing machinery, New York, NY, USA, pp 1–14
  • Elshan, E., Zierau, N., Engel, C. et al. (2022). Understanding the design elements affecting user acceptance of intelligent agents: past, present and future. Inf Syst Front, https://doi.org/10.1007/s10796-021-10230-9
  • Elson, J. S., Derrick, D., & Ligon, G. (2018) Examining trust and reliance in collaborations between humans and automated agents. In: Proceedings of Hawaii international conference on system sciences 2018
  • Fehr, B. (1996). Friendship processes. SAGE.
  • Feine, J., Gnewuch, U., Morana, S., & Maedche, A. (2019). A taxonomy of social cues for conversational agents. International Journal of Human-Computer Studies, S1071581918305238. https://doi.org/10.1016/j.ijhcs.2019.07.009
  • Feine, J., Gnewuch, U., Morana, S., & Maedche, A. (2020). Gender bias in chatbot design. In A. Følstad, T. Araujo, S. Papadopoulos et al (Eds.), Chatbot research and design (pp. 79–93). Springer International Publishing.
  • Fischer-Hübner, S., Angulo, J., Karegar, F., & Pulls, T. (2016). Transparency, privacy and trust – technology for tracking and controlling my data disclosures: Does this work? In S. M. Habib, J. Vassileva, S. Mauw, & M. Mühlhäuser (Eds.), Trust management X (pp. 3–14). Springer International Publishing.
  • Fogg, B. J. (2002). Persuasive technology: Using computers to change what we think and do. Ubiquity, 2002. https://doi.org/10.1145/764008.763957
  • Fox, J., & Gambino, A. (2021). “Relationship Development with Humanoid Social Robots: Applying Interpersonal Theories to Human/Robot Interaction,” _cyberpsychology, Behavior, and Social Networking_, Mary Ann Liebert. Inc., publishers. https://doi.org/10.1089/cyber.2020.0181
  • Gnewuch, U., Morana, S., & Maedche, A. (2017) Towards designing cooperative and social conversational agents for customer service. In: Proceedings of the 38th international conference on information systems (ICIS)
  • Gnewuch, U., Morana, S., Adam, M., & Maedche, A. (2018a) Faster is not always better: understanding the effect of dynamic response delays in human-chatbot interaction. In: proceedings of European conference on information systems 2018
  • Gnewuch, U., Morana, S., Heckmann, C., & Maedche, A. (2018b) Designing conversational agents for energy feedback. In: International conference on design science research in information systems and technology. Springer, pp 18–33
  • Gong, L. (2008). How social is social responses to computers? The function of the degree of anthropomorphism in computer representations. Computers in Human Behavior, 24, 1494–1509. https://doi.org/10.1016/j.chb.2007.05.007
  • Gouldner, A. W. (1960). The norm of reciprocity: A preliminary statement. American Sociological Review, 25, 161–178. https://doi.org/10.2307/2092623
  • Gregor, S. (2002). Design theory in information systems. In Australasian Journal of Information Systems (Vol. Vol. 10). https://doi.org/10.3127/ajis.v10i1.439.
  • Gregor, S. (2006). The nature of theory in information systems. MIS Quarterly, 30(3), 611–642. https://doi.org/10.2307/25148742
  • Gregor, S., & Jones, D. (2007). The anatomy of a design theory. Journal of the Association for Information Systems, 8.
  • Gregor, S., & Hevner, A. R. (2013). Positioning and presenting design science research for maximum impact. MIS Quarterly, 37(2), 337–355. https://doi.org/10.25300/MISQ/2013/37.2.01
  • Gregor, S., Kruse, L. C., & Seidel, S. (2020). Research perspectives: the anatomy of a design principle. Journal of the Association for Information Systems, 21. https://doi.org/10.17705/1jais.00649
  • Guzman, A. (2017). Making AI Safe for Humans: A Conversation With Siri. In Socialbots and Their Friends (pp. 69–82). Routledge.
  • Harvey, P. H., Currie, E., Daryanani, P., & Augusto, J. C. (2016). Enhancing student support with a virtual assistant. In G. Vincenti, A. Bucciero, & C. Vaz de Carvalho (Eds.), E-learning, E-education, and online training (pp. 101–109). Springer International Publishing.
  • Hecht, M. L. (1978). The conceptualization and measurement of interpersonal communication satisfaction. Human Communication Research, 4, 253–264. https://doi.org/10.1111/j.1468-2958.1978.tb00614.x
  • Heerink, M., Kröse, B., Evers, V., & Wielinga, B. (2010). Relating conversational expressiveness to social presence and acceptance of an assistive social robot. Virtual Reality, 14, 77–84. https://doi.org/10.1007/s10055-009-0142-1
  • Hevner, A., March, S. T., Park, J., & Ram, S. (2004). Design science in information systems research. MIS Quarterly, 28, 75–105. https://doi.org/10.2307/25148625
  • Hildebrand, C., & Bergner, A. (2021). Conversational robo advisors as surrogates of trust: Onboarding experience, firm perception, and consumer financial decision making. J of the Acad Mark Sci, 49, 659–676. https://doi.org/10.1007/s11747-020-00753-z
  • Hinde, R. A. (1979). Towards understanding relationships. press. Academic.
  • Hobert, S. (2019) Say hello to ‘coding tutor’! Design and evaluation of a chatbot-based learning system supporting students to learn to program. ICIS 2019 Proceedings
  • Homans, G. C. (1958). Social behavior as exchange. American Journal of Sociology, 63(6), 597–606. https://doi.org/10.1086/222355
  • Horton, D., & Wohl, R. R. (1956). Mass communication and para-social interaction. Psychiatry, 19, 215–229. https://doi.org/10.1080/00332747.1956.11023049
  • Hu, P., Lu, Y., & Gong, Y. (Yale). (2021). Dual humanness and trust in conversational AI: A person-centered approach. Computers in Human Behavior, 119, 106727. https://doi.org/10.1016/j.chb.2021.106727
  • Iivari, J. (2015). Distinguishing and contrasting two strategies for design science research. European Journal of Information Systems, 24(1), 107–115. https://doi.org/10.1057/ejis.2013.35
  • Iivari, J. (2019). A critical look at theories in design science research.
  • Kanda, T., Hirano, T., Eaton, D., & Ishiguro, H. (2004). Interactive robots as social partners and peer tutors for children: A field trial. Human–Computer Interaction, 19, 61–84. https://doi.org/10.1080/07370024.2004.9667340
  • Kankanhalli, A., Tan, B. C. Y., & Wei, K.-K. (2005). Contributing knowledge to electronic knowledge repositories: An empirical investigation. MIS Quarterly, 29, 113–143.
  • Kaplan, A., & Haenlein, M. (2019). Siri, siri, in my hand: Who’s the fairest in the land? On the interpretations, illustrations, and implications of artificial intelligence. Business Horizons, 62, 15–25. https://doi.org/10.1016/j.bushor.2018.08.004
  • Kapp, K. M. (2012). The gamification of learning and instruction. Wiley San Francisco.
  • Kelley, H. H., & Thibaut, J. W. (1978). Interpersonal relations: A theory of interdependence. John Wiley & Sons.
  • Kim, K. J., Park, E., & Shyam Sundar, S. (2013). Caregiving role in human–robot interaction: A study of the mediating effects of perceived benefit and social presence. Computers in Human Behavior, 29, 1799–1806. https://doi.org/10.1016/j.chb.2013.02.009
  • Knijnenburg, B. P., & Willemsen, M. C. (2016). Inferring capabilities of intelligent agents from their external traits. ACM Trans Interact Intell Syst, 6(28), 1–28:25. https://doi.org/10.1145/2963106
  • Krämer, N. C., Eimler, S., von der Pütten, A., & Payr, S. (2011). Theory of companions: What can theoretical models contribute to applications and understanding of human-robot interaction? Applied Artificial Intelligence, 25, 474–502. https://doi.org/10.1080/08839514.2011.587153
  • Krämer, N. C., von der Pütten, A., & Eimler, S. (2012). Human-agent and human-robot interaction theory: Similarities to and differences from human-human interaction. In M. Zacarias & J. V. de Oliveira (Eds.), Human-computer interaction: The agency perspective (pp. 215–240). Springer.
  • Krämer, N. C., Rosenthal-von der Pütten, A. M., & Hoffmann, L. (2015). Social effects of virtual and robot companions. In The handbook of the psychology of communication technology (pp. 137–159). John Wiley & Sons, Ltd.
  • Kuechler, B., & Vaishnavi, V. (2008). On theory development in design science research: Anatomy of a research project. European Journal of Information Systems, 17(5), 489–504. https://doi.org/10.1057/ejis.2008.40
  • Kuechler, W., & Vaishnavi, V. (2012). A framework for theory development in design science research: Multiple perspectives. Journal of the Association for Information Systems, 13(6), 395–423. https://doi.org/10.17705/1jais.00300
  • L’Abbate, M., Thiel, U., & Kamps, T. (2005) Can proactive behavior turn chatterbots into conversational agents? In: IEEE/WIC/ACM international conference on intelligent agent technology. pp 173–179
  • Lankton, N., McKnight, D., & Tripp, J. (2015). Technology, humanness, and trust: Rethinking trust in technology. Journal of the Association for Information Systems, 16. https://doi.org/10.17705/1jais.00411.
  • Lee, S., & Choi, J. (2017). Enhancing user experience with conversational agent for movie recommendation. Int J Hum-Comput Stud, 103, 95–105. https://doi.org/10.1016/j.ijhcs.2017.02.005
  • Lee, S. K., Kavya, P., & Lasser, S. (2021). Social interactions and relationships with an intelligent virtual agent. International Journal of Human-Computer Studies, 102608. https://doi.org/10.1016/j.ijhcs.2021.102608
  • Li, Y., Tee, K. P., Ge, S. S., & Li, H. (2013). Building companionship through human-robot collaboration. In Social Robotics (pp. 1–7). Springer International Publishing.
  • Luger, E., & Sellen, A. (2016) Like having a really bad PA: The gulf between user expectation and experience of conversational agents. In: Proceedings of the 2016 CHI conference on human factors in computing systems. ACM, pp 5286–5297
  • Maedche, A., Morana, S., Schacht, S., Werth, D., & Krumeich, J. (2016). Advanced user assistance systems. Business & Information Systems Engineering, 58(5), 367–370. https://doi.org/10.1007/s12599-016-0444-2
  • March, S. T., & Smith, G. F. (1995). Design and natural science research on information technology. Decision Support Systems, 15(4), 251–266. https://doi.org/10.1016/0167-9236(94)00041-2
  • McCrae, R. R., & John, O. P. (1992). An introduction to the five-factor model and its applications. Journal of Personality, 60(2), 175–215. https://doi.org/10.1111/j.1467-6494.1992.tb00970.x
  • McTear, M., Callejas, Z., & Griol, D. (2016). The conversational interface (Vol. 10, pp. 978). Springer.
  • McTear, M. F. (2017). The rise of the conversational interface: A new kid on the block? Lecture Notes in Computer Science 10341 LNAI, 38–49. https://doi.org/10.1007/978-3-319-69365-1_3
  • McTear, M. (2018) Conversational modelling for ChatBots: Current approaches and future directions. Technical report, Ulster University
  • Mendelson, M. J., & Aboud, F. (2012). McGill friendship questionnaire-respondent’s affection (MFQ-RA). Measurement Instrument Database for the Social Science.
  • Meth, H., Mueller, B., University of Groningen, et al. (2015). Designing a requirement mining system. Journal of the Association for Information Systems, 16, 799–837
  • Miles, M. B., & Huberman, A. M. (1994). Qualitative data analysis: An expanded sourcebook. sage
  • Möller, F., Guggenberger, T. M., & Otto, B. (2020). Towards a Method for Design Principle Development in Information Systems. In S. Hofmann, O. Müller, & M. Rossi (Eds.), Designing for digital transformation. Co-creating services with citizens and industry (pp. 208–220). Springer International Publishing.
  • Moon, Y. (2000). Intimate exchanges: Using computers to elicit self-disclosure from consumers. Journal of Consumer Research, 26(4), 323–339. https://doi.org/10.1086/209566
  • Morana, S., Friemel, C., Gnewuch, U., Maedche, A., & Pfeiffer, J. (2017). Interaktion mit smarten Systemen — Aktueller Stand und zukünftige Entwicklungen im Bereich der Nutzerassistenz. Wirtschaftsinformatik & Management, 9(5), 42–51. https://doi.org/10.1007/s35764-017-0101-7
  • Mori, M. (2012). The uncanny valley: The original essay by Masahiro Mori. In: IEEE Spectrum: Technology, Engineering, and Science News. Accessed 26 Jun 2019. https://spectrum.ieee.org/automaton/robotics/humanoids/the-uncanny-valley
  • Murphy, M. (2019) This app is trying to replicate you. Quartz. Accessed 25 Jan 2022. https://qz.com/1698337/replika-this-app-is-trying-to-replicate-you/
  • Nass, C., Steuer, J., & Tauber, E. R. (1994) Computers are social actors. In: Proceedings of the SIGCHI conference on human factors in computing systems. ACM, pp 72–78
  • Nass, C., Moon, Y., Fogg, B. J., Reeves, B., & Dryer, D. C. (1995). Can computer personalities be human personalities? International Journal of Human-Computer Studies, 43(2), 223–239. https://doi.org/10.1006/ijhc.1995.1042
  • Nass, C., & Moon, Y. (2000). Machines and mindlessness: Social responses to computers. Journal of Social Issues, 56(1), 81–103. https://doi.org/10.1111/0022-4537.00153
  • Nißen, M., Selimi, D., Janssen, A. et al. (2021). See you soon again, chatbot? A design taxonomy to characterize user-chatbot relationships with different time horizons. Computers in Human Behavior:107043. https://doi.org/10.1016/j.chb.2021.107043
  • Park, E., Jin, D., & Del Pobil Ap. (2012). The law of attraction in human-robot interaction. International Journal of Advanced Robotic Systems, 9, 35.
  • Pfeuffer, N., Adam, M., Toutaoui, J. et al. (2019a). Mr. and Mrs. Conversational agent - gender stereotyping in judge-advisor systems and the role of egocentric bias. ICIS 2019 proceedings
  • Pfeuffer, N., Benlian, A., Gimpel, H., & Hinz, O. (2019b). Anthropomorphic information systems. Bus Inf Syst Eng, 61, 523–533.
  • Porcheron, M., Fischer, J. E., Reeves, S., & Sharples, S. (2018). Voice interfaces in everyday life. In: Proceedings of the 2018 CHI conference on human factors in computing systems. association for computing machinery, New York, NY, USA, pp 1–12
  • Premack, D., & Woodruff, G. (1978). Does the chimpanzee have a theory of mind? Behavioral and Brain Sciences, 1, 515–526.
  • Purao, S., Chandra Kruse, L., & Maedche, A. (2020). The origins of design principles: Where do … they all come from?
  • Qiu, L., & Benbasat, I. (2009). Evaluating anthropomorphic product recommendation agents: A social relationship perspective to designing information systems. Journal of Management Information Systems, 25(4), 145–182. https://doi.org/10.2753/MIS0742-1222250405
  • Qiu, L., & Benbasat, I. (2010). A study of demographic embodiments of product recommendation agents in electronic commerce. International Journal of Human-Computer Studies, 68, 669–688. https://doi.org/10.1016/j.ijhcs.2010.05.005
  • Randrup, N., Druckenmiller, D., & Briggs, R. O. (2016) Philosophy of collaboration. In: System sciences (HICSS), 2016 49th Hawaii international conference on. IEEE, pp 898–907
  • Ransbotham, S., Gerbert, P., Reeves, M. et al. (2018). Artificial intelligence in business gets real. MIT Sloan Management Review, 60280.
  • Rawlins, W. (2017). Friendship matters. Routledge.
  • Reeves, S., Porcheron, M., & Fischer, J. (2018). “This is not what we wanted”: Designing for conversation with voice interfaces. Interactions, 26, 46–51. https://doi.org/10.1145/3296699
  • Rheu, M., Shin, J. Y., Peng, W., & Huh-Yoo, J. (2021). Systematic review: Trust-building factors and implications for conversational agent design. International Journal of Human–Computer Interaction, 37, 81–96. https://doi.org/10.1080/10447318.2020.1807710
  • Robert, L. P., Pierce, C., Marquis, L., Kim, S., & Alahmad, R. (2020). Designing fair AI for managing employees in organizations: A review, critique, and design agenda. Human–Computer Interaction, 35(5–6), 545–575. https://doi.org/10.1080/07370024.2020.1735391
  • Rook, K. S. (1987). Social support versus companionship: Effects on life stress, loneliness, and evaluations by others. Journal of Personality and Social Psychology, 52(6), 1132–1147. https://doi.org/10.1037/0022-3514.52.6.1132
  • Rotter, J. B. (1980). Interpersonal trust, trustworthiness, and gullibility. American Psychologist, 35, 1–7.
  • Rzepka, C., & Berger, B. (2018) User interaction with AI-enabled systems: A systematic review of IS research. ICIS 2018 Proceedings
  • Saffarizadeh, K., Boodraj, M., & Alashoor, T. M. (2017). Conversational assistants: Investigating privacy concerns, trust, and self-disclosure. In: ICIS 2017 Proceedings
  • Saldaña, J. (2015). The coding manual for qualitative researchers. SAGE.
  • Sandberg, J., & Alvesson, M. (2021). Meanings of theory: Clarifying theory through typification. Journal of Management Studies, 58, 487–516. https://doi.org/10.1111/joms.12587
  • Schroeder, J., & Schroeder, M. (2018). Trusting in machines: How mode of interaction affects willingness to share personal information with machines. In: Proceedings of the 51st Hawaii international conference on system sciences
  • Seeber, I., Bittner, E., Briggs, R. O. et al. (2019). Machines as teammates: A research Agenda on AI in team collaboration. Information & Management: 103174. https://doi.org/10.1016/j.im.2019.103174
  • Seeger, A.-M., Pfeiffer, J., & Heinzl, A. (2018) Designing anthropomorphic conversational agents: Development and empirical evaluation of a design framework. In: ICIS 2018 Proceedings
  • Seymour, M., Riemer, K., & Kay, J. (2018). Actors, avatars and agents: Potentials and implications of natural face technology for the creation of realistic visual presence. Journal of the Association for Information Systems, 19.
  • Seymour, M., Yuan, L., Dennis, A., & Riemer, K. (2019). Crossing the uncanny valley? Understanding affinity, trustworthiness, and preference for more realistic virtual humans in immersive environments. In: HICSS 2019 Proceedings
  • Seymour, M., Yuan, L., Dennis, A., & Riemer, K. (2021). Have we crossed the uncanny valley? Understanding affinity, trustworthiness, and preference for realistic digital humans in immersive environments. Journal of the Association for Information Systems, 22. https://doi.org/10.17705/1jais.00674.
  • Shah, H., Warwick, K., Vallverdú, J., & Wu, D. (2016). Can machines talk? Comparison of Eliza with modern dialogue systems. Computers in Human Behavior, 58, 278–295. https://doi.org/10.1016/j.chb.2016.01.004
  • Shapiro, S. S., & Wilk, M. B. (1965). An analysis of variance test for normality (complete samples). Biometrika, 52, 591–611.
  • Shawar, B. A., & Atwell, E. (2007). Chatbots: Are they really useful? In: Ldv forum. pp 29–49
  • Siemon, D., Becker, F., Eckardt, L., & Robra-Bissantz, S. (2017). One for all and all for one - towards a framework for collaboration support systems. Education and Information Technologies. https://doi.org/10.1007/s10639-017-9651-9
  • Siemon, D., & Jusmann, S. (2021). Preferred appearance of embodied conversational agents in knowledge management. AMCIS 2021 Proceedings
  • Sin, J., & Munteanu, C. (2020). An empirically grounded sociotechnical perspective on designing virtual agents for older adults. Human–Computer Interaction, 35, 481–510. https://doi.org/10.1080/07370024.2020.1731690
  • Skjuve, M., Følstad, A., Fostervold, K. I., & Brandtzaeg, P. B. (2021). My chatbot companion - A study of human-chatbot relationships. International Journal of Human-Computer Studies, 149, 102601. https://doi.org/10.1016/j.ijhcs.2021.102601
  • Söllner, M., Hoffmann, A., Hoffmann, H., & Leimeister, J. M. (2012). Vertrauensunterstützung für sozio-technische ubiquitäre Systeme. Z Betriebswirtsch, 82(4), 109–140. https://doi.org/10.1007/s11573-012-0584-x
  • Söllner, M., Hoffmann, A., & Leimeister, J. M. (2016). Why different trust relationships matter for information systems users. European Journal of Information Systems, 25, 274–287. https://doi.org/10.1057/ejis.2015.17
  • Solove, D. J. (2004). The digital person: technology and privacy in the information age. NYU Press.
  • Sprecher, S. (1986). The relation between inequity and emotions in close relationships. Social Psychology Quarterly, 49, 309–321. https://doi.org/10.2307/2786770
  • Stever, G. S. (2017). Parasocial theory: Concepts and measures. The International Encyclopedia of Media Effects. American Cancer Society, 1–12.
  • Strohmann, T., Siemon, D., & Robra-Bissantz, S. (2019). Designing virtual in-vehicle assistants: Design guidelines for creating a convincing user experience. AIS Transactions on Human-Computer Interaction, 11, 54–78. https://doi.org/10.17705/1thci.00113
  • Sunyaev, A., Dehling, T., Taylor, P. L., & Mandl, K. D. (2015). Availability and quality of mobile health app privacy policies. Journal of the American Medical Informatics Association, 22(e1), e28–e33. https://doi.org/10.1136/amiajnl-2013-002605
  • Svennevig, J. (2000). Getting acquainted in conversation: A study of initial interactions. John Benjamins Publishing.
  • Ta, V., Griffith, C., Boatfield, C., Wang, X., Civitello, M., Bader, H., DeCero, E., & Loggarakis, A. (2020). User experiences of social support from companion chatbots in everyday contexts: Thematic analysis. Journal of Medical Internet Research, 22(3), e16235. https://doi.org/10.2196/16235
  • Taddei, S., & Contena, B. (2013). Privacy, trust and control: Which relationships with online self-disclosure? Computers in Human Behavior, 29(3), 821–826. https://doi.org/10.1016/j.chb.2012.11.022
  • Tavanapour, N., Poser, M., & Bittner, E. A. C. (2019) Supporting the idea generation process in citizen participation - toward an interactive system with a conversational agent as facilitator. ECIS 2019 Proceedings 18
  • Thuan, N., Drechsler, A., & Antunes, P. (2019). Construction of design science research questions. Communications of the Association for Information Systems, 44. https://doi.org/10.17705/1CAIS.04420
  • Tsiourti, C. (2018). Artificial agents as social companions: Design guidelines for emotional interactions. University of Geneva.
  • Turilli, M., & Floridi, L. (2009). The ethics of information transparency. Ethics and Information Technology, 11(2), 105–112. https://doi.org/10.1007/s10676-009-9187-9
  • Turkle, S. (2010). In good company? Close Engagements with Artificial Companions. John Benjamins, 3–10.
  • Vardoulakis, L. P., Ring, L., Barry, B. et al. (2012). Designing relational agents as long term social companions for older adults. Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) 7502 LNAI:289–302. https://doi.org/10.1007/978-3-642-33197-8-30
  • Vaswani, A., Shazeer, N., Parmar, N. et al. (2017). Attention is all you need. In Advances in neural information processing systems. Curran associates. Inc.
  • Venable, J., Pries-Heje, J., & Baskerville, R. (2016). FEDS: A framework for evaluation in design science research. European Journal of Information Systems, 25(1), 77–89. https://doi.org/10.1057/ejis.2014.36
  • Venkatesh, V., Morris, M. G., Davis, D., & Davis, F. D. User acceptance of information technology: Toward a unified view. (2003). MIS Quarterly, 27(3), 425. 425–478. https://doi.org/10.2307/30036540
  • Venkatesh, V., Brown, S. A., & Bala, H. (2013). Bridging the qualitative-quantitative divide: guidelines for conducting mixed methods research in information systems. MIS Quarterly, 37(1), 21–54. https://doi.org/10.25300/MISQ/2013/37.1.02
  • Veruggio, G., Operto, F., & Bekey, G. (2016). Roboethics: Social and ethical implications. In Springer handbook of robotics (pp. 2135–2160). Springer.
  • Walls, J. G., Widmeyer, G. R., & El Sawy, O. A. (1992). Building an information system design theory for vigilant EIS. Information Systems Research, 3(1), 36–59. https://doi.org/10.1287/isre.3.1.36
  • Walster, E., Walster, G. W., & Berscheid, E. (1978). Equity: Theory and research. https://doi.org/10.1007/BF01542062
  • Wambsganss, T., Janson, A., Soellner, M., & Leimeister, J. M. (2021) AI-based argumentation tutoring – a novel system class to improve learners’ Argumentation Skills. Proceedings 2021:10233. https://doi.org/10.5465/AMBPP.2021.10233abstract
  • Webster, J., & Watson, R. T. (2002). Analyzing the past to prepare for the future: Writing a literature review. MIS Quarterly, xiii–xxiii.
  • Weizenbaum, J. (1966). ELIZA—a computer program for the study of natural language communication between man and machine. Communications of the ACM, 9(1), 36–45. https://doi.org/10.1145/365153.365168
  • Wilks, Y. (2005). Artificial companions. Lecture Notes in Computer Science, 36–45.
  • Winkle, K., Melsión, G. I., McMillan, D., & Leite, I. (2021) Boosting robot credibility and challenging gender norms in responding to abusive behaviour: a case for feminist robots. In: Companion of the 2021 ACM/IEEE international conference on human-robot interaction. Association for computing machinery, New York, NY, USA, pp 29–37
  • Woolf, B. P., Arroyo, I., & Muldner, K. (2010). The effect of motivational learning companions on low achieving students and students with disabilities. In V. Aleven, J. Kay, J. Mostow et al (Eds.), Intelligent tutoring systems (pp. 327–337). Springer.
  • Xiao, J., Stasko, J., & Catrambone, R. (2007) The role of choice and customization on users’ interaction with embodied conversational agents: Effects on perception and performance. In: Proceedings of the SIGCHI conference on human factors in computing systems. Association for computing machinery, New York, NY, USA, pp 1293–1302
  • Young, J. E., Hawkins, R., Sharlin, E., & Igarashi, T. (2008). Toward acceptable domestic robots: Applying insights from social psychology. International Journal of Social Robotics, 1(1), 95. https://doi.org/10.1007/s12369-008-0006-y
  • Yu, H., Shen, Z., Miao, C. et al. (2018). Building ethics into artificial intelligence. In: Proceedings of the 27th international joint conference on artificial intelligence
  • Zhou, L., Gao, J., Li, D., & Shum, H.-Y. (2020). The design and implementation of XiaoIce, an empathetic social chatbot. Computational Linguistics, 46(1), 53–93. https://doi.org/10.1162/coli_a_00368

Appendix A.

Full conversation with sarah and the virtual assistant

Appendix B.

Questionnaire on the virtual classmate instantiation

Appendix C.

Meta-Requirement, design principle and design feature mapping diagram