13,688
Views
8
CrossRef citations to date
0
Altmetric
Research Article

Artificial Intelligence Governance For Businesses

, , ORCID Icon &

ABSTRACT

While artificial intelligence (AI) governance is thoroughly discussed on a philosophical, societal, and regulatory level, few works target companies. We address this gap by deriving a conceptual framework from literature. We decompose AI governance into governance of data, machine learning models, and AI systems along the dimensions of who, what, and how “is governed.” This decomposition enables the evolution of existing governance structures. Novel, business-specific aspects include measuring data value and novel AI governance roles.

Introduction

Artificial Intelligence (AI) has emerged as an essential field in research and practice, with projected spending of almost 100 billion US dollars by 2023 from 38 billion in 2019 (Shirer & Daquila, Citation2019). AI exhibits forms of intelligent behavior, allowing for an extensive range of cost-efficient, well-performing applications. While AI incorporates many techniques, in this work, we understand AI as systems that learn from samples, i.e., these systems rely on models stemming from a subbranch of AI denoted “machine learning.” Machine learning (ML) includes techniques such as deep learning that infer decision-making behavior from data. It is responsible for most successes of AI since learning from data rather than extracting and implementing domain expert rules yields AI systems of superior performance at relatively low costs. For example, AI has shown remarkable success in many application areas (Meske et al., Citation2022), for instance, job recruitment (Pan et al., Citation2022), credit scoring (H. Wang et al., Citation2019), designing floorplans for microchips (Khang, Citation2021), managing predictive maintenance strategies (Arena et al., Citation2022), autopilots in aviation (Garlick, Citation2017), or autonomous driving (Grigorescu et al., Citation2020). Thus, AI is also the focus of many organizations, and research shows that 90% of executives agree that AI is a business opportunity critical to their company’s success (Ransbotham et al., Citation2019). At the same time, only 10% of the executives report a significant financial gain from implementing AI (Ransbotham et al., Citation2020). Consequently, there is still a high level of uncertainty about how AI technology can be used effectively to generate value in organizations and how to specifically leverage the technology to make profits for organizations.

While AI is not new, its recent technical and regulatory advances are staggering (Burt, Citation2021), and they are likely to remain high-paced for some time. Fast advancements in AI make it difficult for companies to keep up and identify appropriate governance mechanisms to reap economic benefits of AI. At the same time, companies must comply with a growing body of regulations touching upon data, ML models and AI systems. Moreover, AI exhibits characteristics that make it both attractive but also challenging to rely on, for instance:

(1) The output of AI is often difficult to understand (Adadi & Berrada, Citation2018), even for applications where its models are easy to build (e.g., using AutoML or adapting existing models). It is hard to comprehend why AI makes a specific decision and how an AI system works in general. The lack of understanding poses challenges to meeting regulatory requirements and improving systems beyond what is already known.

(2) AI produces unexpected results that are partly beyond the control of an organization. It exhibits non-predictable, “ethics”-unaware, data-induced behavior yielding novel security, safety, and fairness issues. AI provides largely statistical guarantees on its performance. AI might malfunction in scenarios that are hard to foresee and exploitable by malicious agents, e.g., as demonstrated for existing self-driving cars (Morgulis et al., Citation2019). An AI model is often trained using objectives such as maximizing overall accuracy and might be based on imbalanced data, e.g., image data that contains more samples of one race than another. This might lead to unethical models that exhibit, for example, a much higher error rate for one race than for another.

(3) Data may lead to biases in decisions of AI systems. Systems with biased decision-making pose a reputation risk as witnessed, e.g., for Amazon’s AI-based recruitment tool (Dastin, Citation2018). This could ultimately lead to legal consequences. While such systems often cause public outcries, they are no surprise. ML models – forming the brain of AI systems – are typically optimized toward a single metric (Goodfellow et al., Citation2016). They commonly focus exclusively on performance, e.g., they maximize the number of correct decisions, ignoring other concerns such as the rationale and information on what these decisions should or should not be based on. This highlights that mechanisms (and governance) to ensure adequate decision-making are needed.

(4) AI is quickly evolving on a technological level, constantly yielding new business opportunities and possibilities for digital innovation. AI components have entered many products and industries that are traditionally unaware of AI technology. It also yields new standards (Cihon, Citation2019) and new regulatory requirements (Burt, Citation2021). While today AI does not exhibit “general intelligence,” it is unknown when it could emerge (Ford, Citation2018). AI might also undermine human authority, which is one reason for likely future regulations, e.g., related to human agency and oversight (European Commission, Citation2020).

Hence, governance mechanisms play an important role to mitigate AI challenges and raise AI potential in organizations. Governance is strongly positively related to firm performance (Bhagat & Bolton, Citation2008). AI governance for business comprises of the structure of rules, practices, and processes used to ensure that the organization’s AI technology sustains and extends the organization’s strategies and objectives. Our goal is to provide a governance perspective for businesses on a tangible level that highlights relevant governance concepts and sets the boundaries and practices for the successful use of AI to meet a company’s objectives, such as profitability and efficiency. At the same time, AI adoption and capabilities are still under research (e.g., Jöhnk et al., Citation2021; Mikalef & Gupta, Citation2021). The focus of this article is not on temporary AI governance during a transient adoption phase but on (long-term) governance of AI. Aside from regulatory aspects, there are also technological contributions to AI that impact governance. They allow governing in novel ways or ways that at least shed some light to what extent governance mechanisms are realistic to implement. In particular, we elaborate on model governance in this work, including testing, data valuation, and data quality issues.

While there is some debate on the disruptiveness of AI on governance (Liu & Maas, Citation2021), current governance structures should be considered both from a research and practitioner perspective. Thus, we approach AI governance from the lens of existing information technology (IT) and data governance frameworks. For instance, there exist models for AI governance, i.e., a layered model for AI governance (Gasser & Almeida, Citation2017), consisting of a social and legal layer (covering norms, regulation, and legislation), an ethical layer (criteria and principles), and a technical layer (data governance, accountability, standards), hence focusing on a strong societal and humanistic aspect. However, existing approaches neglect corporate and practical concerns. With this article, we aim to fill this gap. For this purpose, we derive a conceptual framework by synthesizing the literature on AI, including related fields such as ML.

The paper is structured as follows. First, we provide the methodology and outline the framework. Then we state the outcome of the literature review with details of the outlined framework. Then, we present the entire framework together with a discussion on key differences to other frameworks that originate from the different nature of AI (compared to IT). Next, we provide a discussion and future work. In particular, we show how theory development can be carried out using our framework. We conclude with a summary and avenues for future research.

Methodology

We began our study with a systematic literature search. While this yielded several interesting works, we found them insufficient to cover AI governance from a business perspective, lacking important areas related to governance and AI. Thus, we enhanced the systematic search with a narrative search (King & He, Citation2005) to fill gaps not addressed by articles from the systematic literature review. The narrative search was guided by approaching AI governance from two complementing perspectives (i) looking at governance frameworks on IT (Tiwana et al., Citation2013) and data (Abraham et al. Citation2019) with emphasis on business-relevant aspects and investigating to what extent each aspect in this framework was covered by literature, (ii) looking at key characteristics of AI that are likely impacting their governance and investigating to what extent these characteristics have been addressed. An overview of the process is provided in .

Figure 1. Overview of literature review process.

Figure 1. Overview of literature review process.

We started with a search covering keywords (Machine Learning OR AI OR Artificial Intelligence) AND (governance OR management) AND (compan* OR business* OR firm? OR organization?) on all available databases at our university, i.e., ProQuest, EBSCO, Google Scholar, Science Direct, Springer, and AISnet, restricting to peer-reviewed journals and English articles. Other prominent databases such as Web of Science and Scopus have been superseded by Google Scholar (Martín-Martín et al., Citation2018). The chosen databases also cover our key areas of interest, i.e., managerial works, works in computer science and works at the intersection, e.g., information systems. We considered articles published between 2011 and May 2021. A full-text search gave 30,000 hits on Springer and even more on Google Scholar. Thus, we further limited our search to abstracts. This gave a total of 1,627 hits (1,212 without governance and 415 without management). We estimate that about half of all the hits were duplicates. Articles were removed mostly based on their title and abstract. A large number of articles were non-AI specific (e.g., discussing also blockchain, Industry 4.0 or containing AI, which abbreviated another term) or not focusing on a single company (e.g., public policy articles) or unrelated to governance or management of AI (e.g., using AI to support management). For governance, this left us with 13 out of 415 articles. For management, many articles were domain-specific (e.g., manufacturing) and often discussed specific applications and challenges (e.g., predictive maintenance). While experiences of the management of specific cases are interesting, such articles are of limited relevance for a systematic literature review, leaving us with 17 articles, i.e., a total of 30 articles.

We found that these works on their own did not discuss multiple aspects related to governance as found in the context of IT, data and AI in a business context, e.g., how to value data. They also lacked tangible processes and procedures for ML. The identified works also did not allow to obtain a deeper understanding of key concepts in AI governance, allowing to set practical valid governance policies. Furthermore, a very large body of work is related to practical issues in ML that impact governance, e.g., ML governance and management, testing of ML models, ensuring fairness, explaining “black boxes.” Thus, we also performed a narrative literature review (King & He, Citation2005). A narrative literature review puts the burden of an adequate search procedure on the reviewer. Our search was guided by synthesizing works identified in the systematic literature review and mapping them to existing governance frameworks (Abraham et al., Citation2019; Tiwana et al., Citation2013), taking into account key characteristics of AI such as its autonomy, ability to learn and obscurity. Parts of the framework that were not covered were explicitly searched for. Thus, we used a broad range of keywords, e.g., “AI design,” “AI best practices,” “AI safety,” “AI strategy,” “AI policy,” “AI standards,” “AI performance measurement,” “data quality,” “data valuation,” “ML management,” “ML process,” “Model governance,” “ML testing” accompanied by forward and backward search (Webster & Watson, Citation2002). Since the narrative literature review includes many search terms, we relied on a single engine. We used Google Scholar as it was shown to supersede prominent databases like Web of Science and Scopus (Martín-Martín et al., Citation2018). We removed the limitation to journals since new ideas are often presented first at academic conferences, and a significant body of works, particularly in computer science, only appear as conference articles. We focused on surveys if literature was abundant (for a keyword or during forward search). If there was a lack of peer-reviewed work, we also included gray literature, i.e., primary literature in the form of preprints from arxiv.org and reviews from MIT and Harvard. Other disciplines, namely computer science, frequently cite preprints on arxiv.org, which applies a moderation process for basic quality assurance, but no peer-review system. Pre-prints were only considered after our quality assessment. The main motivation for our process is (1) to include the most recent knowledge for the reader and (2) to cover the broad field of AI and governance from different angles. We believe this approach is necessary to ensure practical relevance and adequately discuss future research opportunities accounting for research that is likely completed soon. That is to say, pre-prints are typically scientific works currently or very soon under-peer review. Neglecting them puts a reader at risk of focusing on research gaps that are already in the process of being closed.

To analyze the identified works, we leverage existing frameworks on IT and data governance (as described in section Framework Outline). We used core, more generic dimensions shown in . Individual works identified in the review process were first assigned into more generic categories shown in . This was possible for all works. The generic categories contained sub-categories that stemmed from existing frameworks but were continuously expanded in the review process. We formed new categories for works that did not fit into the framework. Our final framework also dismisses aspects of prior frameworks (on IT and data governance) if they are not relevant in our context, as elaborated in the discussion of our final framework.

Figure 2. Our perspective on AI governance (left) and the core of our conceptual framework (right).

Figure 2. Our perspective on AI governance (left) and the core of our conceptual framework (right).

While most of the literature analysis was done by one researcher, the resulting framework was discussed by all coauthors.

Framework outline

We first describe the two existing frameworks that built the foundation of our framework. We defer a detailed discussion of differences to the end of the paper, presenting the final framework ().

Abraham et al. (Citation2019) presented a framework for data governance. It covered only the data dimension, which we expand with concepts that arise or allow for different, more specific treatments in the context of AI such as data valuation, bias, and (concept) drift. For data governance, we focus on differences in governance resulting from a company using data for AI or other purposes. For the latter, we refer to Abraham et al. (Citation2019) instead. The model and system dimensions are of equal importance to the data dimension. The dimensions and concepts are largely a synthesis of multiple existing works stemming from non-technical articles on AI governance and more technical articles on ML. We also state antecedents and consequences. Antecedents describe internal or external factors that impact AI governance. These include internal factors that are prevalent during the adoption of AI governance, as well as internal and external factors that are at best marginally influenced by governance initiatives such as legal requirements. Consequences describe outcomes of AI governance that can be directly or indirectly related to a corporation’s goals. provides an overview of the research framework. , toward the end of the article, includes additional details.

Abraham et al. (Citation2019) built upon Tiwana et al. (Citation2013), which addresses the who, what, and how of IT governance. In this work, for the who dimension, we primarily emphasize the firm and project dimension, only touching upon the governance of large units of analysis such as ecosystems. We recognize that AI being a composition of data, models, and systems, lends itself naturally to govern both IT artifacts (models, systems manifested as software code) and content (data) found in the “what dimension” of Tiwana et al. (Citation2013). As shown in the following sections, we refine Abraham et al. (Citation2019) by defining AI governance as the union of data, model, and systems governance (see, ).

Literature review

This section elaborates on AI governance based on the reviewed papers. First, we define the scope in the form of an AI governance definition more clearly before employing the structure of the conceptual framework shown in . Each dimension of the conceptual framework is refined with findings from literature. We start with the description of governance mechanisms and afterward move on to subject scope, organizational scope, and targets.

AI governance definition

To clarify the scope of this work, we state a working definition of the key term “AI governance for Businesses.” It combines the well-defined term “corporate governance” and “AI,” which shows more variance in the literature. We propose the following definition:

AI governance for business is the structure of rules, practices, and processes used to ensure that the organization’s AI technology sustains and extends the organization’s strategies and objectives.

Aligned with Abraham et al. (Citation2019), we emphasize the following six parts of AI governance:

  • fostering collaboration across functions

  • structuring and formalizing AI management through a framework

  • focusing on AI as a strategic asset

  • defining how and who makes decisions

  • developing supporting artifacts (policy, standards, and procedures), and

  • monitoring compliance.

AI aims at building “machines that can compute how to act effectively and safely in a wide variety of novel situations” (Russell & Norvig, Citation2020). ML can be seen as a branch of AI with the narrower goal of learning from data. Thus, ML denotes the models and processes for building and testing models by learning from data. AI aims at building an AI system consisting of an ML model and typically other components, e.g., graphical user interface and input handling for an AI (software) system or physical elements such as those found for a robot. The ML model might be seen as the brain of the AI, i.e., it is responsible for making decisions.

The term AI differs from the one of ML. It is not only broader regarding its goal but also in terms of the techniques used to achieve intelligent behavior. While ML focuses on learning from data, AI includes non-learning-based techniques, e.g., logical deduction (Russell & Norvig, Citation2020). We use a data-driven lens of AI based on the observation that most existing AI techniques, e.g., deep learning, incorporate some form of learning from data given as a set of samples.

Data representing experience in the form of samples might be based on an exploration of an environment by an agent using reinforcement learning (RL; Russell & Norvig, Citation2020) like in “Alpha Go Zero” (Silver et al., Citation2017). In this case, no domain knowledge in the form of data is available, but the agent generates it through exploration. We do not discuss RL in detail, mainly because businesses reap most benefits with supervised models and RL is less commonly used in practice (Witten & Frank, Citation2017).

From a governance perspective, it is appealing and natural to define AI in the following three areas: data, model, and (AI-)system since they relate to existing governance areas. Furthermore, they build on top of each other. AI governance refers to data and IT governance, covering both systems (commonly a large software code basis) and models (typically a small code basis). In contrast, model governance is a new area that comprises the governance of relatively small software codes (compared to traditional software) that contain models and procedures for training, evaluation, and testing.

Governance mechanisms

Businesses can make use of various mechanisms to govern AI. They consist of formal structures connecting business, IT, data, model, ML, and system management functions, formal processes and procedures for decision-making and monitoring, as well as practices supporting the active participation and collaboration among stakeholders (De Haes & Van Grembergen, Citation2009; Peterson, Citation2004). Following the literature on IT governance (De Haes & Van Grembergen, Citation2009; Peterson, Citation2004), we employ the separation between (a) structural, (b) procedural, and (c) relational governance mechanisms. An overview is given in .

Figure 3. Overview of governance mechanisms.

Figure 3. Overview of governance mechanisms.

Structural mechanisms

Structural governance mechanisms define reporting structures, governance bodies, and accountability (Borgman et al., Citation2016). They comprise (1) roles and responsibilities and (2) the allocation of decision-making authority. While for data governance, there is a rich set of literature (Abraham et al., Citation2019), literature is relatively sparse for AI governance. Ho et al. (Citation2019) briefly discussed the responsibilities of a chief data officer (CDO) and chief information officer (CIO) in the context of image analysis in healthcare. Using a hub that centralizes responsibilities such as talent recruitment, performance management, and AI standards, processes, and policies has also been set forth (Fountaine et al., Citation2019). In the adoption phase, establishing a center of excellence has been suggested (Kruhse-Lehtonen & Hofmann, Citation2020).

Developing AI systems is a multi-disciplinary endeavor (Tarafdar et al., Citation2019). Thus, establishing an (interdisciplinary) AI governance council that has been advocated for AI in healthcare (Reddy et al., Citation2020) might be needed to handle the complex interrelation between model outputs, training data, regulatory and business requirements. Executive sponsors also play an essential role (Pumplun et al., Citation2019). To what extent the executive sponsor controls and decides on performance goals might depend on the company’s stage of adoption. For example, Pumplun et al. (Citation2019) advocated that at least in the early phase of adoption, a dedicated AI budget without any performance obligations might benefit adoption. More specific roles related to model aspects are still subject to investigation. Serban et al. (Citation2020) view the assignment of an owner to each feature as best practice.

Procedural mechanisms

Procedural governance mechanisms aim to ensure that AI systems and ML models operate correctly and efficiently, they are held securely, and their operation meets legal and company internal requirements and policies concerning explainability, fairness, accountability, security, and safety. While ultimately, these goals must be fulfilled for the entire system and the ML model, explainability, fairness, and accountability are primary concerns of ML models. In contrast, safety is a more holistic trait, e.g., while an ML model might be deemed unsafe, safeguards such as non-AI-based backup components providing only reduced functionality might ensure the safety of the entire system. A “big red emergency button” might at least allow disabling the AI at any time, though the usefulness of this idea is subject to debate (Arnold & Scheutz, Citation2018). Procedural mechanisms also aim to ensure data, model, and system-relevant traits and targets, as shown in . For data, model and systems individually and holistically, they comprise (1) a strategy; (2) policies; (3) standards; (4) processes and procedures; (5) contractual agreements; (6) performance measurement; (7) compliance monitoring; and (8) issue management.

The strategy denotes high-level guidance of actions based on strategic business objectives. The review by Keding (Citation2020) investigated the interplay of AI and strategic management, focusing on two aspects (a) antecedents, i.e., data-driven workflows (data value chain and data quality), managerial willingness, and organizational determinants (AI strategy and implementation), (b) consequences of AI in strategic management on an individual and organizational level (e.g., human-AI collaboration, AI in business models). The work is focused on the use of AI for decision-making within a company, i.e., within a business intelligence context. There is work on several elements related to AI strategy, such as guiding principles by academics and practitioners (Iansiti & Lakhani, Citation2020; Kruhse-Lehtonen & Hofmann, Citation2020; Pichai, Citation2018; Smit & Zoet, Citation2020).

AI policies provide high-level guidelines and rules. Organizations use AI policies to communicate key objectives, accountability, roles, and responsibilities. There is active discussion in politics on AI policies. For example, in a white paper by the European Union (EU; European Commission, Citation2020), policy options have been set forth on how not to hinder the uptake of AI while limiting risks, and a subset of it is already part of a proposal (European Commission, Citation2021). On an organizational level, one source for policies might be best practices (Alsheiabni et al., Citation2020). Measurement in (public) AI policy is also an active field of debate and research, e.g., Mishra et al. (Citation2020) discuss aspects of how to measure AI’s societal and economic impact and risks and threats of AI systems.

While the development of standards concerning data has widely progressed (DAMA International, Citation2009), the development of AI standards is currently ongoing, e.g., Wellbeing metrics for ethical AI (IEEE P7010), benchmarking accuracy of facial recognition systems (IEEE P7013), fail-safe design for AI systems (IEEE P7009), ethically driven AI Nudging methodologies (IEEE P7008) (see, Cihon (Citation2019) for a more comprehensive overview). In our framework, standards should ensure that data representations, ML models, and the architecture of AI systems, including their accompanying processes, are consistent and normalized throughout the organization. They should support interoperability within and across organizations and ensure their fitness for purpose (Mosley et al., Citation2010).

According to International Organization for Standardization’s (ISO) Standard 9000, a process is a “set of interrelated or interacting activities, which transforms inputs into outputs” and procedures as a “specified way to carry out an activity or a process.” In the context of governance, we view procedures and processes as standardized, documented, and repeatable methods. AI processes can be considered a fundamental element of a successful AI governance implementation. They relate to data governance processes (Abraham et al., Citation2019) and processes tied to ML and AI systems. ML model development relies commonly on elements of the cross-industry standard process for data mining (CRISP-DM; Wirth & Hipp, Citation2000). Coding guidelines in software engineering can also serve as a governance mechanism. Similarly, developing and enforcing model (and system) architectural principles, documentation, and coding guidelines with the goal of better composability and maintainability might also be a governance instrument for AI to foster interoperability (Stoica et al., Citation2017). Serban et al. (Citation2020) view the documentation of each feature, including its rationale, as best practice.

Other proposals have been made with influences from software engineering (Amershi et al., Citation2019; Liu et al., Citation2020; Rahman et al., Citation2019; Watanabe et al., Citation2019). While Amershi et al. (Citation2019) rely on surveys within a single company, Rahman et al. (Citation2019) generalize from a case study describing the use of ML to correct transaction errors. However, there is still active discussion on which process is best (Liu et al., Citation2020). Palczewska et al. (Citation2013) discussed model governance listing processes for model evaluation (e.g., review policies), control (changes and security, documentation), and validation. Testing procedures of ML systems received significant attention (Zhang et al., Citation2020), as well as assurance of ML (Ashmore et al., Citation2021; Hamada et al., Citation2020), i.e., executing an ML development process that delivers evidence for the model’s safety. Furthermore, some of the works covered under processes might also exhibit a procedural character. Overall, there is a lack of research on model and AI procedures.

Contractual agreements between participating internal departments or external organizations might be related to data but also models and AI systems. Models contain information on the training data that might be extracted and abused in various ways, e.g., by competitors to reduce costs for data labeling. For example, consider a system that allows uploading an image and obtaining a classification. A competitor might use the system to label her unlabeled data. The competitor might use the resulting dataset to train a novel classifier and offer an “identical” service. Contracts also play an important role in reducing liability risk by clearly specifying operational boundaries of AI. AI might malfunction under circumstances that are non-intuitive and unexpected for a human. Therefore, operational constraints and limitations require explicit legal notices. Classical risk management approaches have also been suggested for AI (Clarke, Citation2019). AI agency risk and the mitigation using business process management have been discussed on a conceptual level (Sidorova & Rafiee, Citation2019).

Compliance monitoring tracks and enforces conformance with regulatory requirements and organizational policies, standards, procedures, and contractual agreements. Prominent regulations include the European General Data Protection Regulation (GDPR), which touches upon data and models, e.g., it grants the right to explanations of model decisions. GDPR is supported by extensive guidelines on compliance (Privacy Team, Citation2020). While there is some conceptual understanding of AI from a legal perspective (Scherer, Citation2015), and there are multiple non-legal statements from governments, e.g., European Commission (Citation2020), the body of regulations and, thus, the need for compliance monitoring is likely to grow. Audits allow for interrogating complex processes to clarify whether these processes adhere to company policy, industry standards, or regulations. Brundage et al. (Citation2020) discuss AI auditing. They emphasize its need due to the complexity of AI.

Issue management refers to identifying, managing, and resolving AI-related issues. For data issues (Abraham et al., Citation2019), this includes processes for the standardization of data issues and their resolution, the nominating persons for issue resolution, as well as an escalation process. For model and AI systems, little information is available.

Relational mechanisms

Relational governance mechanisms facilitate collaboration between stakeholders. They encompass (1) communication, (2) training, and (3) the coordination of decision-making. Serban et al. (Citation2020) express the need to communicate within an interdisciplinary ML team and suggest utilizing a collaborative development platform. Training procedures for data scientists have been put forward in Biomedicine by Van Horn et al. (Citation2018). They analyzed a rich set of training resources and conducted training events on applied projects. Mikalef and Gupta (Citation2021) stated that technical and business skills are essential AI capabilities. Thus, training employees is critical. While training might often refer to developing skills to leverage AI, it might also refer to training employees whose tasks and responsibilities might be automated or augmented by AI to mitigate negative impacts (Kaplan & Haenlein, Citation2019). Communication might help reduce employees’ fear, e.g., by signaling a company’s intention to use AI as augmentation of workers rather than a replacement (Kaplan & Haenlein, Citation2019). Overall, little is known for ML models and AI systems governance in contrast to data governance (Abraham et al., Citation2019; Borgman et al., Citation2016).

Subject scope

This dimension pertains to what is governed. It covers IT artifacts (models and systems) and content (data), as shown in .

Figure 4. Overview of subject scope.

Figure 4. Overview of subject scope.

Data subject

Data is the representation of facts using text, numbers, images, sound, or video (DAMA International, Citation2009). While Abraham et al. (Citation2019) provide an overview of the data scope, we emphasize that the type of data impacts the model selection, e.g., structured or tabular data vs. unstructured data such as text and images. For example, deep learning models work well with large unstructured data but have been relatively less successful for tabular data. An essential characteristic of data is its sensitivity level: Is it personal or non-personal data? Personal data, i.e., data that relates to humans, underlies different regulations, e.g., GDPR enacted in the European Union in 2018. A secondary reason for the separation is that human data often come with different characteristics with respect to quality and costs. These are strongly reflected in ML and system properties. For instance, GDPR grants the right to explanation to individuals for automated decisions based on their data. Such explainability processes are less of a concern for models based on non-personal data. Reddy et al. (Citation2020) set forth a governance model for healthcare based on fairness, transparency, trustworthiness, and accountability. Janssen et al. (Citation2020) emphasized the importance of a trusted data-sharing framework.

Model subject

Supervised learning uses labeled training data. That is, each sample X is associated with a label y. Unsupervised learning consists of unlabeled training data. It typically addresses tasks like organizing data into groups (clustering) or generating new samples. While adaptivity might be a requirement in a dynamic environment, online systems, which continuously learn from observed data, are more complex from a system engineering perspective with multiple unresolved challenges (Stoica et al., Citation2017). System evolution must be foreseen and managed, which is not needed for offline systems that do not change. In the past, such dynamic systems have gotten “out of control” after deployment, e.g., Microsoft’s Tay chatbot had to be shut down after only a few hours of operation due to inappropriate behavior (Mason, Citation2016). Finally, an important concept is that of transparency of ML models. Transparent models are intrinsically human understandable, whereas complex black-box models such as deep learning require external methods that provide explanations that might or might not suffice to understand the model (Adadi & Berrada, Citation2018; Meske et al., Citation2022; Schneider & Handali, Citation2019).

System subject

Autonomy refers to the decision-power of AI. It is commonly discussed in the context of agents (Franklin & Graesser, Citation1996). Autonomy impacts legal responsibilities and liability. The discussion of autonomy relates closely to human-AI collaboration. For example, a fully autonomous AI might decide on the treatment of a patient without any human review or possibility of intervention. At the same time, an assisting AI might only suggest a treatment, and the decision is left to a human. AI might augment jobs or tasks, requiring close interaction with human collaborators or fully automate them (Davenport, Thomas & Kirby, Citation2015). There are widespread concerns about the future of work in the face of AI (Jarrahi, Citation2018) and policy recommendations by regulatory bodies on the designation of decision control (European Commission, Citation2020). Human-AI collaboration also plays an important role from a strategic perspective (Keding, Citation2020). A survey of human-AI interaction (Rzepka & Berger, Citation2018) categorized works into the topic of users, system, interaction, task, context, and outcomes. Relevant system characteristics also cover the appearance of AI and gestural and conversational behavior. These characteristics also play a prominent role in assisting digital AI agents (Maedche et al., Citation2019) and social robots (Nocentini et al., Citation2019). Guidelines for human-AI interaction have also been set forth by Amershi et al. (Citation2019). The guidelines serve for initial usage, interaction, behavior in case of errors, and behavior over time. They can also be categorized as informing users, managing decision control, and collecting and leveraging user feedback. AI safety refers to the property of not causing danger, risk, or injury (Amodei et al., Citation2016). Highly safety-critical systems typically put human lives at risk, such as autonomous cars. The environment in which an AI system operates strongly impacts its properties’ relevance and required levels, such as security and robustness. Dynamic environments are subject to rapid and large changes, while static environments do not evolve. Dynamic environments often require online learning, i.e., highly adaptive models that continuously learn.

Organizational scope

The organizational scope defines the breadth of governance in terms of entities involved. Following Tiwana et al. (Citation2013), it consists of the intra- and inter-organizational scope as shown in . The intra-organizational scope refers to governance within a single organization whereas the inter-organizational scope refers to governance between multiple organizations. For AI governance in an intra-organizational scope, this relates to quality and integrity of data, models, and systems on a project level. At the same time, the firm-level considers the demands of multiple stakeholders dispersed within the company. While these aspects are well-studied for data governance (Abraham et al., Citation2019), more research is needed for model and system governance. Co-development among multiple partners, including public entities such as universities or private entities such as individuals in hackathons, are seen as ways for successful AI development (Kaplan & Haenlein, Citation2019). Such collaborations require appropriate governance. Aside from direct co-development, existing knowledge or artifacts such as models are often freely available, e.g., open-source deep learning modes or open-access research papers. Models might be trained using publicly available data and code or even be pre-trained on a large dataset and only adapted for a company’s specific need, e.g., using transfer learning on a small, company-owned dataset. While open source initiatives and platforms are not new, AI arguably further pushed the extent of these initiatives, e.g., most designs of deep learning models for image tasks or natural language processing (NLP) originate from publicly available models. They are often obtained from repositories such as GitHub or already integrated into frameworks such as Tensorflow or Pytorch. Governance of open-source code or platforms is a topic on its own (Halckenhaeusser et al., Citation2020). Their governance poses new challenges for the governance of AI in organizations and raises questions how to ensure that such models are compliant with corporate policies and how to mitigate liability issues. This is particularly relevant because models might suffer from different forms of biases and might malfunction in non-predictable ways. Also, their inner workings are hard to assess due to their black-box nature.

Figure 5. Overview of organizational scope.

Figure 5. Overview of organizational scope.

Targets

Governance programs address goals in one or multiple areas, as summarized in . While we discuss data, models, and systems separately, most works discuss only one aspect. Siebert et al. (Citation2020) propose guidelines for assessing qualities of ML systems, discussing criteria related to data, model, environment, infrastructure, and system. We subsume their model criteria focusing on metrics relevant to governance.

Figure 6. Overview of targets.

Figure 6. Overview of targets.

Data targets

Data governance aims to maximize data value at minimal cost (Abraham et al., Citation2019). However, data valuation is currently not part of data governance and AI (Abraham et al., Citation2019). The value of data is difficult to estimate since potential uses of data and the resulting benefits are hard to foresee. Pricing data on data marketplaces or quantifying the right amount of data to purchase is non-trivial (Jiao et al., Citation2018; Pei, Citation2020). For example, if data comes for free and is not traded, e.g., in the form of trace data associated with business transactions, its value is often not of primary concern due to low storage costs and data usage only inside the business. However, data valuation gains in relevance if the acquisition of data comes with costs, e.g., if data has to be labeled by humans as part of the construction of a dataset for an AI system or if data requires costly processing, such as manual cleansing to raise data quality. Some general principles qualitatively describe the value of data for ML, such as that the value of data increases with quality and exhibits diminishing marginal utility. Some works allow quantifying the impact of a data sample on the correctness of an ML model, e.g., using XAI techniques (Ghorbani & Zou, Citation2019; Meske et al., Citation2022) though they are often computationally demanding. These techniques answer the question: “How much accuracy is lost (or gained) if we do not utilize a specific data sample for training a model?.” This provides a solid basis for data valuation of single samples but also datasets. More proactively, active learning (Aggarwal, Citation2015) allows for determining data samples that are giving most improvements for a given model. In ML, training data serves the only purpose to increase model performance. The value of a performance increase of an ML model might often be better and easier to quantify than that of data in general (without knowing any context on its usage). For example, improvements can be valued by increasing revenue from recommendations for a recommender system in an online shopping setting. This allows to more easily assess the value of data for a particular system – down to a single sample. Quantifying data value at least on an aggregate level is highly desirable to make informed investment decisions. The value of data also changes over time. While there might be multiple reasons for this, concept drift is a major concern in the field of AI that reduces data value.

Concept drift refers to how the data distribution changes over time (Gama et al., Citation2014; Lu et al., Citation2018). Concept drift implies that the data used to train the ML model does not capture the relationship that the model should capture. Concept drift typically gets more severe over time. One can distinguish between real and virtual drift (Lu et al., Citation2018). Real drift refers to a change in the classification. It captures the scenario that a particular action might be deemed legal up to a regulatory change and, from then on, illegal. Virtual drift denotes a change in distribution of samples between training and operation. For example, a change of distribution might occur if a system used and designed for adults is used increasingly by younger people in somewhat different ways. Janssen et al. (Citation2020) proposed monitoring the outcomes of algorithms. However, ideally, changes in data are detected before algorithms are impacted, as discussed in literature on concept drift, e.g., Zhang et al. (Citation2020).

Data quality denotes the ability of data to meet its usage requirements in a given context (Khatri & Brown, Citation2010). The relevance and handling of data quality might differ for AI systems compared to business intelligence applications or other forms of analytics. In the context of AI, better data quality is always preferable from a model performance perspective. However, improving data quality also comes with high costs. Given that state-of-the-art ML models have shown some robustness to certain forms of data quality issues (Frénay & Verleysen, Citation2013; Nigam et al., Citation2020; Song et al., Citation2020), the preferred data quality from an economic perspective might be a revenue tradeoff between increased model performance and costs for data cleansing. There also exist techniques to filter “bad” data that might even deteriorate model performance when used for training – see, Brodley and Friedl (Citation1999) for pioneering work and Ghorbani and Zou (Citation2019) for a recent technique. Therefore, data filtering techniques might replace data cleansing given that a large amount of mostly accurate data exists.

Another data quality aspect related to fairness of ML systems is bias. Data is biased if it is not representative of the population or phenomenon of study. Biased data might lead to biased (“unfair”) ML models (Verma & Rubin, Citation2018) that violate regulations.

Data lineage includes tracking its origin, where it moves and how it is processed. In particular, in AI, this includes connecting data, pre-processing, and models using the pre-processed data (for training; Vartak et al., Citation2016). The same data might be used in many different ML models, e.g., through transfer learning. Thus, proper governance should ensure that any updates, e.g., improvements in data quality, should be propagated adequately to all relevant stakeholders.

The data lifecycle represents the approach of defining, collecting, creating, using, maintaining, archiving, and deleting data (e.g., Khatri & Brown, Citation2010).

This step further encompasses deriving data retention requirements from business, regulatory, and accountability demands.

Model targets

Similarities and differences of data and model targets can be seen by comparing the targets for data and model in . Model quality relates to a number of different concepts, which might be extended and less relevant depending on the application. Security and privacy relate to two separate matters: Security is a target of a model, e.g., “Is the model secure against attacks during operation?” and an attribute related to handling the model, e.g., “Is the model stored securely with proper access rights (in the development environment)?” For data, only the second concern applies. While metadata is highly important to establish the semantics of data, for models, this is typically encoded in the artifact itself, i.e., the model definition. However, metadata on models such as who authored or changed the model is relevant to ensure accountability.

ML models are commonly assessed and tested for the following dimensions (Zhang et al., Citation2020): correctness, efficiency, fairness, interpretability, robustness, security, and privacy. Zhang et al. (Citation2020) also list model relevance, which measures to what extent a model overfits or underfits the data as testing property. However, it is implicit that relevance of a model increases with correctness of the model.

Correctness of an ML model measures the probability that the ML system works as intended. For classification tasks, it is often reported as the percentage of correctly categorized samples of a classifier (accuracy) for a test dataset. It is often seen as the key determinant for model performance judgment. Robustness defines to what extent an ML model can function correctly in the presence of invalid inputs (Zhang et al., Citation2020). For instance, it allows assessing to what extent concept drift is a concern. It is also related to security since common attacks on ML models exploit their non-robustness. More generally, security is the resilience against negative impacts caused by manipulations or illegal access to ML models. Privacy, which is also a security concern, relates to information confidentiality. It measures to what extent an ML system can maintain private data, e.g., if private data be extracted from a model trained on private data.

Decisions of an ML model might not be in accordance with accepted or legally binding “fairness” definitions. There is no consensus on what fairness comprises, and multiple, potentially even conflicting definitions exist (Verma & Rubin, Citation2018). However, there is widespread acceptance among the public and legal authorities that decision criteria denoted as protected characteristics (Corbett-Davies & Goel, Citation2018) such as gender, religion, familial status, age, and race must not be used. While fairness might often be traced back to data, it should primarily be assessed in ML models for two reasons: (1) decisions are made by models (not by data); (2) the impact of data bias on the fairness of decisions of an ML model is difficult to anticipate. Data might be biased, and a model’s decisions are not (that much) or the other way around. For illustration, there are methods to counteract fairness issues on a model level (Du et al., Citation2020), i.e., a fair model might be obtained using such mechanisms even in the face of data bias.

Explainability relates to understanding model behavior. A common definition states that it provides the degree to which a model’s decision or the model itself can be understood. Understanding might improve trust of end-users as well as help in model testing and validation by engineers. Efforts of explainable artificial intelligence (XAI) are also driven by legislation, e.g., GDPR granting rights to individuals for an explanation involving personal data. XAI is still subject to research within the information systems community (Adadi & Berrada, Citation2018; Meske et al., Citation2022; Schneider & Handali, Citation2019) and other disciplines (Das & Rad, Citation2020). Efficiency of an ML system relates to the computational effort to construct a model or perform a prediction. In a practical setting, it is often approximated by the time it takes a computer to complete either of the two aforementioned tasks, i.e., model construction and prediction. It is also a key focus of computer science, often estimating worst-case behavior measured in abstract operations. In particular, if AI systems are to be deployed on resource-constrained systems such as mobile devices, efficiency of prediction is a primary concern.

In contrast, efficiency of model construction might limit productivity of data scientists and ML engineers and impact overall project duration. ML models might take weeks to train on ordinary personal computers, particularly if multiple models should be assessed. In contrast, distributed training using specialized hardware (Verbraeken et al., Citation2020) of the very same model might take only a few minutes. Thus, infrastructure relates to the software and hardware available to train, test and deploy ML models. Enabling parallel training of ML is considered a best practice (Serban et al., Citation2020). It relates to efficiency and scalability of AI systems to a growing number of predictions as well as to growing complexity and a growing number of models. While there are differences in hardware used, e.g., GPUs and scheduling jobs for ML (Mayer & Jacobsen, Citation2020), overall governance aspects are similar to non-AI specialized IT.

According to Sridhar et al. (Citation2018), solutions for production ML should track provenance/lineage, ensure reproducibility, enable audits and compliance checks of models, foster reusability, handle scale and heterogeneity, allow for flexible metadata usage. Sridhar et al. (Citation2018) go beyond these recommendations and provide an architecture and design to facilitate “model governance.” The complex nature of AI has even led to the idea of employing third-party audits on internal development (as a common routine; Brundage et al., Citation2020).

Model management is an emerging topic (Vartak et al., Citation2016), with an increasing number of tools becoming available, e.g., Chen et al. (Citation2020) provide a recent example. While it is often developer-focused, it also addresses points relevant to model governance, managing modeling experiments, enabling reproducibility, and supporting sharing of models. Model management comes with challenges on a conceptual, data management and engineering level (Schelter et al., Citation2018). These challenges include reproducibility issues, deciding when to retrain a model, lack of abstraction for the whole ML pipeline (such as the Structured Query Language [SQL] for data), access to model metadata, multi-language codebase. Challenges of ML management have also been expressed relative to software management leading to three main challenges (Amershi et al., Citation2019): (1) discovering, managing, and versioning is more complex for ML (than for data or software only), (2) model customization, and reuse require skills not found in software teams and (3) AI components are more difficult to be handled as separate models due to their entanglement and error behavior. A model management system is supposed to track the data pipeline, i.e., to record which data and transformations have been used to build and test a model. This is analogous to versioning using code repositories for classical software. It should also support the sharing of models and, potentially, levels of abstractions of ML frameworks.

Model management systems might furthermore support testing, monitoring, and analysis of models. Serban et al. (Citation2020) promote software engineering best practices for ML grouped into six classes, i.e., data, training, coding, deployment, team, and governance. Several practices relate to model management, recommended testing practices. Palczewska et al. (Citation2013) discussed model governance based on data governance tailored toward predictive models for toxicology in food and related industries. While they miss some aspects of contemporary model targets such as fairness and explainability, they discuss an information system for model management.

System targets

The overall system is commonly assessed using performance metrics that strongly depend on model performance and encompasses other metrics such as system usability. Systems that collaborate with humans are of particular interest. They pursue different goals, e.g., the collaboration of AI and humans can either substitute or augment existing work or assemble new dynamic configurations of humans and AI that provide a single, integrated unit (Rai et al., Citation2019). For systems interacting with humans, data guardianship, system accountability, and ethics in decision-making as well as security and privacy need to be considered (Rai et al., Citation2019). For example, a company might pursue the objective to minimize risks due to accountability resulting from AI decisions by augmenting user tasks rather than substituting, as is observed in a slow transitioning from assisted driving to autonomous driving (Bellet et al., Citation2019). AI safety (Hernández-Orallo et al., Citation2019) exhibits a rich set of characteristics, e.g., describing an operator’s ability to intervene during operation, disposition to self-modify (autonomy), explainability of goals and behavior. It also relates to model robustness.

Furthermore, the system ultimately makes decisions based on model outputs. Thus, security and safety mechanisms such as a “big red emergency button” to disable the AI can be incorporated on a system level. AI ethics and implications have been extensively discussed in academic literature (e.g., Floridi et al. (Citation2018)), and increasingly by institutional bodies (e.g., European Union (2020)). AI governance can aim to realize AI opportunities from a societal perspective, such as cultivating social cohesion, increasing societal capabilities, and limiting risks such as devaluing human skills and reducing human control (Floridi et al., Citation2018).

Antecedents

Antecedents are the external and internal factors that influence the adoption of governance practices (Tallon et al., Citation2013, p. 143). We mainly synthesize factors of data governance from Abraham et al. (Citation2019). We also investigated literature that discusses factors related to AI adoption (Chui & Malhotra, Citation2018; Fountaine et al., Citation2019), readiness (Jöhnk et al., Citation2021) and capability (Mikalef & Gupta, Citation2021) as well as antecedents specific to the use of AI-as-a-service (Zhang et al., Citation2020). We added existing AI development procedures and AI capability as (internal) antecedents based on these works. The rationale is that if a company has good development procedures and high AI capability, it must handle many concerns relevant to AI governance, such as data and model qualities. As such, different governance mechanisms might depend on AI capability. In particular, AI-business alignment might be relevant. It relates to its readiness based on harmonized and AI-aligned processes related to decision-making (Jöhnk et al., Citation2021). Furthermore, governance might also consider existing AI (development) procedures. Other internal factors relate to strategy, resources, and culture. Common factors such as top management support and financial budget also play a vital role (Abraham et al., Citation2019; Jöhnk et al., Citation2021). IT in general and enterprise architecture specifically have also been shown to be relevant as antecedents for data governance and AI adoption (Stecher et al., Citation2020). Since models build on data, collecting or accessing large amounts of high-quality data is essential. Cultural aspects relate to innovativeness, willingness to change, risk appetite, and collaborative work (Gupta and Mikalef, Citation2021; Jöhnk et al., Citation2021). Collaboration emphasizes the integration of different perspectives such as business domain (data), model and systems, and skillsets (technical and business) across organizational units.

External factors constitute legal and regulatory requirements relating to data such as GDPR (Privacy Team, Citation2020) or models and systems, e.g., effectively enforcing human agency for self-driving cars due to liability concerns. While AI is seen as a disruptive general-purpose technology (Magistretti et al., Citation2019), current AI is also deemed “narrow AI” due to its success for specific tasks only, i.e., its inability to show general intelligence. Thus, not surprisingly, AI deployment is industry and application-dependent (Chui & Malhotra, Citation2018). Customers must be ready for AI, and they must understand, e.g., its limitations, such as lack of transparency (Jöhnk et al., Citation2021). Customers are one of the stakeholders to be governed. As such, their readiness impacts governance.

Consequences

We categorized the consequences of AI governance by extending Abraham et al. (Citation2019) and Tiwana et al. (Citation2013) into performance effects, risk management, orchestration, and, additionally, effects on people. It is worthwhile to emphasize effects on people due to the widespread discussion of AI and human interaction within the organization, i.e., employees and AI, and of the organization with outsiders, i.e., customers and AI systems of the organization.

It is also possible to dissect the impact of AI according to the subject to be governed, e.g., data, model, system, and stakeholder. Abraham et al. (Citation2019) presented benefits of data governance on performance and risk management. Brundage et al. (Citation2020) presented data governance intending to achieve trust. AI governance aims to increase AI capabilities derived from resources such as data, technology, human skills, and organizational aspects (Gupta et al., Citation2021). AI capabilities have contributed positively to firm performance and creativity (Gupta et al., Citation2021; Amabile, Citation2020). Furthermore, the use of AI (as a service) has also been shown to have a positive impact on firm performance (Zapadka et al., Citation2020).

An outcome of AI governance on employees, e.g., by prescribing adequate training for employees, might be the mitigation of negative automation effects. Impacted performance outcomes could be profitability, growth, stability, efficiency, or survival of a company. Organizational properties that are probably impacted by governance include efficiency, effectiveness, and flexibility. Even more than IT governance, AI governance is likely impacting orchestration, e.g., innovation partitioning. AI competencies and business knowledge are commonly separated, e.g., among departments, platform ecosystems, or cloud providers (offering ML services). Both entities being specialized differently might contribute to innovation.

Final framework

Our work treated AI governance based on existing IT and data governance knowledge, seemingly viewing it as an evolution thereof. In particular, we used the IT governance cube of Tiwana et al. (Citation2013), grounding our framework on the three general dimensions relating to what, who, and how is governed. While this seems to confirm the validity of the existing framework, we argue for an updated framework depicted in and a more refined form in .

Figure 7. AI governance cube based on the IT governance cube (Tiwana et al., Citation2013). it can serve as a framework for theory building through rotation and permutation.

Figure 7. AI governance cube based on the IT governance cube (Tiwana et al., Citation2013). it can serve as a framework for theory building through rotation and permutation.

Figure 8. Concepts within the conceptual framework for data governance. concepts.

Figure 8. Concepts within the conceptual framework for data governance. concepts.

Tiwana’s distinction of IT artifacts, content, and stakeholders seems only partially adequate. The notion of a model does not match the classical notion of an IT artifact, a piece of software or hardware that often contains data or information. A model is more declarative, and data determine its behavior. The exact content of a trained model is difficult to retrieve due to the black-box nature of AI, and it can only be said to be some function of the data. For many organizations, software and hardware are purchased externally and are fairly static. That is, updates are rare and costly since software must be explicitly programmed. In contrast, AI might be self-learning, and better-performing models might be obtained and serve primarily internal purposes. From our perspective IT artifacts as defined by Tiwana in terms of hardware and software might be seen as infrastructure that serves to store and process data as well as to develop models and, potentially, deploy AI systems. We view data, models, and systems as very tightly coupled, though they do not have to be governed as one unit, i.e., they might be governed using different targets.

Furthermore, Tiwana et al. considered the IT architecture as a potential means to govern. While enterprise architecture is relevant for AI adoption (Stecher et al., Citation2020), architectures, once implemented, are difficult to change, making it difficult to use a governance instrument in a dynamic environment. In the context of models, a model’s architecture is relevant. Still, the objective they optimize and the training data are also highly relevant, making the term architecture unsuitable as an overarching term. Therefore, we exclude architecture. In the context of AI, aside from structural mechanisms (relating to decision rights) and procedural mechanisms (relating to control and standards and practices), relational mechanisms such as coordination of decision-making and training are highly important, which are not prominently present in Tiwana’s cube. These three types of mechanisms originate from Peterson (Citation2004) and De Haes and Van Grembergen (Citation2009).. Concepts within the conceptual framework for data governance. Concepts from (Abraham et al., Citation2019) are in dashed boxes with concepts in blue and italics.

Discussion and future work

This work views AI as systems learning from provided data and decomposed AI artifacts to be governed into data, models, and systems. Models differ from software implementing explicit human stated rules, since models derive their behavior in an opaque manner from data and require special XAI techniques to (partially) understand. They also pose potential ethical and security risks, e.g., due to biased decision-making and adversarial attacks. At the same time models are very dynamic and can easily be updated or might even evolve autonomously by (re)training on newly collected data. These factors combined with growing legislation pose challenges for businesses to achieve the goal of leveraging the full potential of AI while mitigating risks. We believe that thorough governance is a key aspect of accomplishing this goal. In our governance cube that can guide companies and researchers alike, AI artifacts (data, model and systems) are part of the “What is governed?” dimension that together with two additional dimensions, i.e., “How is governed?” and “Who is governed?” as well as antecedents and consequences define our framework for AI governance. Tiwana et al. (Citation2013) proposed theory development through permutation (and rotation) in their IT governance cube. Scholars were advised to pick a cell of the cube and study a selection of antecedents and consequences. We also call for doing so for our AI governance cube since we have presented various new mechanisms for governance and highlighted differences in antecedents and consequences compared to IT governance.

Many cells of the AI governance cube require more research and as AI is quickly evolving adjustments to governance are also likely needed. The future of AI might also bring a shift from current “narrow” AI toward artificial general intelligence (AGI). AGI will probably be even more autonomous optimizing pre-specified goals without being given data. That is, it might derive its behavior by learning from her own collected data (rather than relying on given data) like agents in reinforcement learning as seen in ground-breaking systems like AlphaGo. This might bring about new challenges, such as the correct specification of goals and, in particular, adequate constraining of “allowed” solutions to achieve these goals. For example, ethical behavior must be ensured, i.e., by forbidding AI to deceive or endanger people to accomplish a goal. We hope and believe that despite the unpredictable evolution of AI, our framework will endure the test of time with some possible extensions. For example, we believe that data, models and systems will remain key AI artifacts, possibly enhanced by other artifacts such as the aforementioned specification of goals and constraints for AI. That is, even if data is collected and generated automatically, it is still of value and relevant for system understanding, e.g., the costs for training AlphaGo by learning from 29 Mio. played games are estimated to be more than 30 Mio. US dollars.Footnote1

Future developments of AI might shift the focus of governance – in particular, related to security, safety, and human-AI collaboration. Also, concerns with respect to ethics that are only briefly discussed in this work might become even more practically relevant. Already today, many aspects of AI governance remain non-specified and require more attention. While we highlighted many of these throughout the manuscript, we aim to contribute using several practical proposals, whose validity has yet to be verified:

The first set of proposals relates to structural mechanisms, i.e., roles. We propose to mimic the general structure of data governance (Abraham et al., Citation2019), however, with some alterations based on the following rationale: Both data and models share some characteristics that bear relevance for governance. Both should be shareable and employable in various areas of a business (Serban et al., Citation2020). Both undergo a lifecycle (often coupled). Both might be subject to audits and require accountability and testing. They also differ in various ways: Most production quality models undergo an engineering, training, and tuning process, which requires extensive technical knowledge. Data undergoes no processing aside from comparably simple transformations (cleansing, pre-processing) and computing KPIs, e.g., for data quality assessment. Data might also be used for AI unrelated systems. We suggest also specifying separate roles for ML development and ML (and AI) testing. This is also done for software to reduce the risk of bugs. Due to the prominent nature of AI risks, one might also introduce an AI risk manager that might, for example, veto against the rollout of an AI system.

With respect to infrastructure, we propose separating IT infrastructure for developing, testing, and deployment AI systems and ML systems. The reasoning is similar for software systems. Minimizing the number of people by carefully choosing access rights who can get a hold of sensitive data on a production server improves security. It reduces, e.g., the risk of privacy violations of customers. Separating systems also ensures that issues during development that might lead to outages of development infrastructure (e.g., running out of memory, utilizing all CPUs and GPUs) do not impact availability of deployed services. Furthermore, development hardware might be chosen to be of lower quality (and therefore be cheaper) since (short) outages are typically less of a concern as for a production environment.

Furthermore, future work might also focus on stakeholders such as employees and customers, which pertain to a special role, which could be studied in depth on its own. For once, they interact with an AI system. Therefore, they are exposed to risks associated with AI, e.g., customers might be exposed to unexpected behavior for inputs resulting in danger or, if AI is used for automation, it might cause job loss or require increased emphasis on continual learning to cope with change (Kaplan & Haenlein, Citation2019). Employees and customers might also provide data used for training models, often creating tension between their rights (e.g., related to data protection) and a business’s intention to collect their valuable data. They might also have reservations due to trust issues (Kaplan & Haenlein, Citation2019). Furthermore, the development of models and AI systems often involves an interdisciplinary team requiring both technical and business skills (Gupta & Mikalef Citation2021). At least for some time to come, governments and society alike show a high level of interest in AI, making them stakeholders. They contribute to the discussion on ethics that is also ongoing by both academics and non-academics (Almeida et al., Citation2020; European Commission, Citation2020; Pagallo et al., Citation2019; Y. Wang et al., Citation2020; Winfield et al., Citation2019). Governments take an active role in monitoring AI and setting forth new regulations. Hurley (Citation2018) focuses on how the US Department of Defense uses AI to better position itself for cyber events and challenges in the future. Their Responsible, Accountable, Consulted, and Informed (RACI) framework seeks to reconcile the imbalance between bringing in all relevant stakeholders and achieving organizational agility. Thus, while there is some work on stakeholders, a comprehensive view is missing. This follows, in particular, when investigating the governance cube in , calling for structural, procedural, and relational mechanisms for stakeholder governance.

Furthermore, we also elaborated on the novel concept of governing “data value” on a more quantitative level, discussing costs, data quality, and model performance. These concepts also deserve more attention. Existing techniques provide a solid basis to assess the impact of data samples on model performance, which might serve as a surrogate for data value. But they lack predictive capability to answer questions such as “How much will our model (and data) be worth based on anticipated model performance?” Concept drift might render data worthless, but extant works primarily focus on reacting to concept drift rather than mitigating it proactively. Thus, it seems difficult to determine when data (or models) should be sold, updated, or acquired based on forecasts of their future performance. Even if data is not subject to concept drift, it is still likely to devalue over time, e.g., due to improved sensor capabilities such as cameras that capture higher resolution images that allow training ML models that perform better on specific object detection tasks.

Conclusion

This study illuminated the building blocks of AI governance. Our work reviews existing literature on AI and ML using a conceptual framework. While most areas in AI governance deserve more attention in general, our review highlights areas that seem particularly interesting from a business perspective. Aside from a conceptual framework, we also provided more fine-granular views and proposals motivated by technological developments, e.g., related to data quality and value, and analogies to related disciplines, e.g., roles in software development and data governance. Future work can expand on these proposals, e.g., investigating governance roles and data value leveraging our AI governance cube providing a variety of options by picking a specific cell along the three dimensions “What is governed?,” “How is it governed?” and “Who is governed?”

Disclosure statement

No potential conflict of interest was reported by the author(s).

Additional information

Notes on contributors

Johannes Schneider

Johannes Schneider is an Assistant Professor in Data Science at the University of Liechtenstein. He has a Ph.D. in Computer Science from the Eidgenoessische Technische Hochschule (ETH) Zurich. Dr. Schneider’s research has been published in journals such as European Journal of Information Systems (EJIS), Transactions on Software Engineering, Information Systems Management (ISM) and Data Mining and Knowledge Discovery(DMKD), and conference proceedings such as Foundations of Computer Science (FOCS), International Joint Conferences on Artificial Intelligence (IJCAI), Conference on Information and Knowledge Management (CIKM). His research interests encompass technical and managerial aspects of Artificial Intelligence and Data Science.

Rene Abraham

Rene Abraham is a Ph.D. candidate at the University of Liechtenstein. He studied Computer Science at Technische Universität Darmstadt and gathered work experience in the financial services industry. His research interest comprises data governance in intra-organizational and inter-organizational settings.

Christian Meske

Christian Meske is a Full Professor of Socio-technical System Design and Artificial Intelligence at the Institute of Work Science and Faculty of Mechanical Engineering at Ruhr-Universität Bochum, Germany. His research on the design and management of digital innovations has been published in journals such as Business & Information Systems Engineering (BISE), Business Process Management Journal (BPMJ), Communications of the Association for Information Systems (CAIS), International Journal of Information Management (IJIM), Information Systems Frontiers (ISF), Information Systems Journal (ISJ), Information Systems Management (ISM), or MIS Quarterly Executive (MISQE).

Jan Vom Brocke

Jan vom Brocke is the Hilti Endowed Chair of Business Process Management and Director of the Institute of Information Systems at the University of Liechtenstein. His work has been published in, among others, in Management Science, MIS Quarterly, Information Systems Research, Journal of Management Information Systems, Journal of the Association for Information Systems, European Journal of Information Systems, Information Systems Journal, Journal of Information Technology, and MIT Sloan Management Review.

Notes

1. This can be computed from Alpha Go’s paper (Silver et al., Citation2017) as done in https://www.yuzeh.com/data/agz-cost.html

References

  • Abraham, R., Schneider, J., & Vom Brocke, J. (2019). Data governance: A conceptual framework, structured review, and research agenda. International Journal of Information Management, 49, 424–438. 10.1016/j.ijinfomgt.2019.07.008
  • Adadi, A., & Berrada, M. (2018). Peeking inside the black-box: A survey on explainable Artificial Intelligence (XAI). IEEE Access, 6, 52138–52160. 10.1109/ACCESS.2018.2870052
  • Aggarwal, C. C. (2015). Data mining: The textbook. Springer.
  • Almeida, P., Santos, C., & Farias, J. S. (2020). Artificial intelligence regulation: A meta-framework for formulation and governance. In Proceedings Of The 53rd Hawaii International Conference On System Sciences, 10.24251/hicss.2020.647
  • Alsheiabni, S., Messom, C., Cheung, Y., & Alhosni, M. (2020). Winning AI strategy: six-steps to create value from artificial intelligence. In Proceedings of the Americas Conference on Information Systems.
  • Amabile, T. M. (2020). Creativity, artificial intelligence, and a world of surprises. Academy of Management Discoveries, 6(3), 351–354. doi:10.5465/amd.2019.0075
  • Amershi, S., Weld, D., Vorvoreanu, M., Fourney, A., Nushi, B., Collisson, P. et al. (2019). Guidelines for human-ai interaction. In Proceedings of the CHI Conference on Human Factors in Computing Systems, 1–13. 10.1145/3290605.3300233
  • Amodei, D., Olah, C., Steinhardt, J., Christiano, P., Schulman, J., & Mané, D. (2016). Concrete problems in AI safety. arXiv Preprint arXiv, 1606.06565
  • Arena, S., Florian, E., Zennaro, I., Orrù, P. F., & Sgarbossa, F. (2022). A novel decision support system for managing predictive maintenance strategies based on machine learning approaches. Safety Science, 146, Article 105529. 10.1016/j.ssci.2021.105529
  • Arnold, T., & Scheutz, M. (2018). The “big red button” is too late: An alternative model for the ethical evaluation of ai systems. Ethics and Information Technology, 20(1), 59–69. 10.1007/s10676-018-9447-7
  • Ashmore, R., Calinescu, R., & Paterson, C. (2021). Assuring the machine learning lifecycle: Desiderata, methods, and challenges. ACM Computing Surveys (CSUR). doi:10.1145/3453444
  • Bellet, T., Cunneen, M., Mullins, M., Murphy, F., Pütz, F., Spickermann, F., Braendle, C., & Baumann, M. F. (2019). From semi to fully autonomous vehicles: New emerging risks and ethico-legal challenges for human-machine interactions. Transportation Research. Part F, Traffic Psychology and Behaviour, 63, 153–164. doi:10.1016/j.trf.2019.04.004
  • Bhagat, S., & Bolton, B. (2008). Corporate governance and firm performance. Journal of Corporate Finance, 14(3), 257–273. 10.1016/j.jcorpfin.2008.03.006
  • Borgman, H., Heier, H., Bahli, B., & Boekamp, T. (2016). Dotting the I and Crossing (out) the T in IT governance: new challenges for information governance. In Hawaii International Conference on System Sciences (HICSS), (pp. 4901–4909). 10.1109/hicss.2016.608
  • Brodley, C. E., & Friedl, M. A. (1999). Identifying mislabeled training data. Journal of Artificial Intelligence Research, 11, 11. 131–167. 10.1613/jair.606
  • Brundage, M., Avin, S., Wang, J., Belfield, H., Krueger, G., Hadfield, G. et al. (2020). Toward trustworthy AI development: Mechanisms for supporting verifiable claims. arXiv Preprint arXiv, 2004.07213
  • Burt, A. (2021). New AI regulations are coming. is your organization ready? Harvard Business Review
  • Chen, A., Chow, A., Davidson, A., DCunha, A., Ghodsi, A., Hong, S. A. et al. (2020). Developments in MLflow: A system to accelerate the machine learning lifecycle. In Proceedings of the International Workshop on Data Management for End-to-End Machine Learning, (pp. 1–4). 10.1145/3399579.3399867
  • Chui, M., & Malhotra, S., (2018), “AI adoption advances, but foundational barriers remain”Mckinsey and company. Retrieved May 01, 2021, from https://www.mckinsey.com/featured-insights/artificial-intelligence/ai-adoption-advances-but-foundational-barriers-remain
  • Cihon, P. (2019). Standards for AI governance: International standards to enable global coordination in ai research & development. In Future of Humanity Institute. University of Oxford.
  • Clarke, R. (2019, August). Principles and business processes for responsible AI. Computer Law & Security Review, 35(4), 410–422
  • Corbett-Davies, S., & Goel, S. (2018). The measure and mismeasure of fairness: A critical review of fair machine learning. arXiv preprint arXiv, 1808.00023
  • DAMA International. (2009). Dama guide to the data management body of knowledge (1st edition ed.). Technics Publications.
  • Das, A., & Rad, P. (2020). Opportunities and challenges in explainable artificial intelligence (XAI): A survey. arXiv preprint arXiv, 2006. 11371
  • Dastin, J. (2018). Amazon scraps secret ai recruiting tool that showed bias against women. https://www.reuters.com/article/us-amazon-com-jobs-automation-insight/amazon-scraps-secret-ai-recruiting-tool-that-showed-bias-against-women-idUSKCN1MK08G
  • Davenport, Thomas, H., & Kirby, J. (2015). Beyond automation. Harvard Business Review, 59–65
  • De Haes, S., & Van Grembergen, W. (2009). An exploratory study into IT governance implementations and its impact on business/IT alignment. Information Systems Management, 26(2), 123–137. 10.1080/10580530902794786
  • Du, M., Yang, F., Zou, N., & Hu, X. (2020). Fairness in deep learning: A computational perspective. IEEE Intelligent Systems. 36. 4. doi:10.1109/MIS.2020.3000681
  • European Commission. (2020). White paper on artificial intelligence. A European approach to excellence and trust., Retrieved November 02, 2020, from 10.2139/ssrn.3706099
  • European Commission. (2021). Laying down harmonized rules on artificial intelligence Retrieved May 01, 2021, from https://eur-lex.europa.eu/legal-content/EN/TXT/?uri=CELEX%3A52021PC0206
  • Floridi, L., Cowls, J., Beltrametti, M., Chatila, R., Chazerand, P., Dignum, V., Luetge, C., Madelin, R., Pagallo, U., Rossi, F., Schafer, B., Valcke, P., & Vayena, E. (2018). AI4People—an ethical framework for a good AI society: Opportunities, risks, principles, and recommendations. Minds and Machines, 28(4), 689–707. doi:10.1007/s11023-018-9482-5
  • Ford, M. (2018). Architects of intelligence: The truth about ai from the people building it. Packt Publishing Ltd.
  • Fountaine, T., McCarthy, B., & Saleh, T. (2019). Building the ai-powered organization. Retrieved November 03, 2020, from https://hbr.org/2019/07/building-the-ai-powered-organization
  • Franklin, S., & Graesser, A. (1996). Is it an agent, or just a program?: A taxonomy for autonomous agents. In International Workshop On Agent Theories, Architectures, And Languages, (pp. 21–35). 10.1007/bfb0013570
  • Frénay, B., & Verleysen, M. (2013). Classification in the presence of label noise: A survey. IEEE Transactions on Neural Networks and Learning Systems, 25(5), 845–869
  • Gama, J., Žliobaitė, I., Bifet, A., Pechenizkiy, M., & Bouchachia, A. (2014). A survey on concept drift adaptation. ACM Computing Surveys (CSUR), 46(4), 1–37. 10.1145/2523813
  • Garlick, B. (2017). Flying smarter: AI & machine learning in aviation autopilot systems. Retrieved November 04, 2020, from https://mse238blog.stanford.edu/2017/08/bgarlick/flying-smarter-ai-machine-learning-in-aviation-autopilot-systems/
  • Gasser, U., & Almeida, V. A. (2017). A layered model for AI governance. IEEE Internet Computing, 21(6), 58–62. 10.1109/MIC.2017.4180835
  • Ghorbani, A., & Zou, J. (2019). Data shapley: equitable valuation of data for machine learning. Proceedings of the 36th International Conference on Machine Learning.
  • Goodfellow, I., Bengio, Y., Courville, A., & Bengio, Y. (2016). Deep learning (Vol. 1, pp. No. 2). MIT press Cambridge.
  • Grigorescu, S., Trasnea, B., Cocias, T., & Macesanu, G. (2020). A survey of deep learning techniques for autonomous driving. Journal of Field Robotics, 37. 3. 1–25. doi:10.1002/rob.21918
  • Halckenhaeusser, A., Foerderer, J., & Heinzl, A. (2020). Platform governance mechanisms: An integrated literature review and research directions. In Proceedings of the European Conference on Information Systems (ECIS).
  • Hamada, K., Ishikawa, F., Masuda, S., Matsuya, M., & Ujita, Y. (2020). Guidelines for quality assurance of machine learning-based artificial intelligence. In SEKE2020: the 32nd International Conference on Software Engineering & Knowledge Engineering (pp. 335–341). 10.1142/s0218194020400227
  • Hernández-Orallo, J., Martínez-Plumed, F., Avin, S., & Heigeartaigh, S. O. (2019). Surveying safety-relevant ai characteristics. In AAAI workshop on Artificial Intelligence Safety (SafeAI 2019) (pp. 1–9).
  • Ho, C., Soon, D., Caals, K., & Kapur, J. (2019). Governance of automated image analysis and artificial intelligence analytics in healthcare. Clinical Radiology, 74(5), 329–337. 10.1016/j.crad.2019.02.005
  • Hurley, J. (2018). Enabling successful artificial intelligence implementation in the department of defense. Journal of Information Warfare, 17(2), 65–82
  • Iansiti, M., & Lakhani, K. R. (2020). Competing in the age of ai: Strategy and leadership when algorithms and networks run the world. Harvard Business Press. doi:10.1080/09700161.2021.1918951
  • Janssen, M., Brous, P., Estevez, E., Barbosa, L. S., & Janowski, T. (2020). Data governance: Organizing data for trustworthy Artificial Intelligence. Government Information Quarterly, 37(3), 101493.
  • Jarrahi, M. H. (2018). Artificial intelligence and the future of work: Human-ai symbiosis in organizational decision making. Business Horizons, 61(4), 577–586. 10.1016/j.bushor.2018.03.007
  • Jiao, Y., Wang, P., Feng, S., & Niyato, D. (2018). Profit maximization mechanism and data management for data analytics services. IEEE Internet of Things Journal, 5(3), 2001–2014. 10.1109/JIOT.2018.2819706
  • Jöhnk, J., Weißert, M., & Wyrtki, K. (2021). Ready or not, AI comes—An interview study of organizational ai readiness factors. Business & Information Systems Engineering, 63(1), 5–20.
  • Kaplan, A., & Haenlein, M. (2019). Siri, Siri, in my hand: Who’s the fairest in the land? On the interpretations, illustrations, and implications of artificial intelligence. Business Horizons, 62(1), 15–25. 10.1016/j.bushor.2018.08.004
  • Keding, C. (2020). Understanding the interplay of artificial intelligence and strategic management: Four decades of research in review. Management Review Quarterly, 71. 3. 1–44. doi:10.1007/s11301-020-00181-x
  • Khang, A. B. (2021). Superhuman floorplans for microchips. Nature, 594, (27). 207–212. doi:10.1038/d41586-021-01515-9
  • Khatri, V., & Brown, C. V. (2010). Designing data governance. Communications of the ACM, 53(1), 148–152. 10.1145/1629175.1629210
  • King, W. R., & He, J. (2005). Understanding the role and methods of metaanalysis in is research. Communications of the Association for Information Systems, 16(1), 32. 10.17705/1CAIS.01632
  • Kruhse-Lehtonen, U., & Hofmann, D. (2020). How to define and execute your data and AI strategy. Retrieved 2020, November 01, from https://hdsr.mitpress.mit.edu/pub/4vlrf0x2/release/2
  • Liu, H., Eksmo, S., Risberg, J., & Hebig, R. (2020). Emerging and changing tasks in the development process for machine learning systems. In Proceedings of The International Conference on Software and System Processes (pp. 125–134). 10.1145/3379177.3388905
  • Liu, H. Y., & Maas, M. M. (2021). ‘Solving for X?’ Towards a problem-finding framework to ground long-term governance strategies for artificial intelligence. Futures, 126, 102672.
  • Lu, J., Liu, A., Dong, F., Gu, F., Gama, J., & Zhang, G. (2018). Learning under concept drift: A review. IEEE Transactions on Knowledge and Data Engineering, 31(12), 2346–2363. doi:10.1109/tkde.2018.2876857
  • Maedche, A., Legner, C., Benlian, A., Berger, B., Gimpel, H., Hess, T., Hinz, O., Morana, S., & Söllner, M. (2019). Ai-based digital assistants. Business & Information Systems Engineering, 61(4), 535–544. doi:10.1007/s12599-019-00600-8
  • Magistretti, S., Dell’Era, C., & Petruzzelli, A. M. (2019). How intelligent is Watson? Enabling digital transformation through artificial intelligence. Business Horizons, 62(6), 819–829. 10.1016/j.bushor.2019.08.004
  • Martín-Martín, A., Orduna-Malea, E., Thelwall, M., & López-Cózar, E. D. (2018). Google scholar, web of science, and Scopus: A systematic comparison of citations in 252 subject categories. Journal of Informetrics, 12(4), 1160–1177. 10.1016/j.joi.2018.09.002
  • Mason, P. (2016). The racist hijacking of Microsoft’s chatbot shows how the internet teems with hate. Retrieved November 01, 2020, from https://www.theguardian.com/world/2016/mar/29/microsoft-tay-tweets-antisemitic-racism
  • Mayer, R., & Jacobsen, H. A. (2020). Scalable deep learning on distributed infrastructures: Challenges, techniques, and tools. ACM Computing Surveys (CSUR), 53(1), 1–37
  • Meske, C., Bunde, E., Schneider, J., & Gersch, M. (2022). Explainable Artificial Intelligence: objectives, stakeholders and future research opportunities. Information Systems Management, 39(1), 53–63
  • Mikalef, P., & Gupta, M. (2021). Artificial intelligence capability: conceptualization, measurement calibration, and empirical study on its impact on organizational creativity and firm performance. Information & Management, 58(1), 103434. doi:10.1016/j.im.2021.103434
  • Mishra, S., Clark, J., & Perrault, C. R. (2020). Measurement in AI Policy: opportunities and challenges. arXiv, 2009.09071 [cs].
  • Morgulis, N., Kreines, A., Mendelowitz, S., & Weisglass, Y. (2019). Fooling a real car with adversarial traffic signs. arXiv Preprint arXiv, 1907.00374.
  • Mosley, M., Brackett, M., Earley, S., & Henderson, D. (2010). The DMA Guide to the Data Management Body of Knowledge (1st ed.). Technics Publications LLC.
  • Nigam, N., Dutta, T., & Gupta, H. P. (2020). Impact of noisy labels in learning techniques: A survey. Advances in Data and Information Sciences, 403–411. doi:10.1007/978-981-15-0694-9_38
  • Nocentini, O., Fiorini, L., Acerbi, G., Sorrentino, A., Mancioppi, G., & Cavallo, F. (2019). A survey of behavioral models for social robots. Robotics, 8(3), 54. 10.3390/robotics8030054
  • Pagallo, U., Aurucci, P., Casanovas, P., Chatila, R., Chazerand, P. D., Dignum, V., Luetge, C., Madelin, R, Schafer, B., Valcke, P. (2019). Ai4people-on good AI governance: 14 priority actions, a smart model of governance, and a regulatory toolbox. Retrieved October 30, 2020, from https://www.eismd.eu/wp-content/uploads/2019/11/AI4Peoples-Report-on-Good-AI-Governance_compressed.pdf
  • Palczewska, A., Fu, X., Trundle, P., Yang, L., Neagu, D., Ridley, M., & Travis, K. (2013). Towards model governance in predictive toxicology. International Journal of Information Management, 33(3), 567–582. 10.1016/j.ijinfomgt.2013.02.005
  • Pan, Y., Froese, F., Liu, N., Hu, Y., & Ye, M. (2022). The adoption of artificial intelligence in employee recruitment: The influence of contextual factors. The International Journal of Human Resource Management, 33(6), 1125–1147. 10.1080/09585192.2021.1879206
  • Pei, J. (2020). A survey on data pricing: From economics to data science. IEEE Transactions on Knowledge and Data Engineering. 10.1109/tkde.2020.3045927
  • Peterson, R. R. (2004). Integration strategies and tactics for information technology governance. Strategies for Information Technology Governance, 37–80. doi:10.4018/978-1-59140-140-7.ch002
  • Pichai, S. C. (2018). AI at Google: Our principles. Retrieved November 04, 2020, from https://www.blog.google/technology/ai/ai-principles
  • Privacy Team, I. G. (2020). EU general data protection regulation (GDPR)–an implementation and compliance guide. IT Governance Ltd. doi:10.2307/j.ctv17f12pc
  • Pumplun, L., Tauchert, C., & Heidt, M. (2019). A new organizational chassis for artificial intelligence-exploring organizational readiness factors. In European Conference on Information Systems (ECIS).
  • Rahman, M. S., Rivera, E., Khomh, F., Guéhéneuc, Y.-G., & Lehnert, B. (2019). Machine learning software engineering in practice: An industrial case study. arXiv preprint arXiv, 1906.07154.
  • Rai, A., Constantinides, P., & Sarker, S. (2019). Next generation digital platforms:: toward human-AI hybrids. MIS Quarterly, 43(1). 3–4
  • Ransbotham, S., Khodabandeh, S., Fehling, R., LaFountain, B., & Kiron, D. (2019). Winning with ai. Retrieved November 01, 2020, from https://sloanreview.mit.edu/projects/winning-with-ai/
  • Ransbotham, S., Khodabandeh, S., Kiron, D., Candelon, F., Chu, M., & LaFountain, B. (2020). Are you making the most of your relationship with AI? Retrieved November, 01, 2020, from https://www.bcg.com/publications/2020/is-your-company-embracing-full-potential-of-artificial-intelligence
  • Reddy, S., Allan, S., Coghlan, S., & Cooper, P. (2020). A governance model for the application of AI in health care. Journal of the American Medical Informatics Association, 27(3), 491–497. 10.1093/jamia/ocz192
  • Russell, S., & Norvig, P. (2020). Artificial intelligence: A modern approach (4th edition. ed.). Pearson.
  • Rzepka, C., & Berger, B. (2018). User interaction with ai-enabled systems: A systematic review of is research. In International Conference on Information Systems (ICIS)
  • Schelter, S., Biessmann, F., Januschowski, T., Salinas, D., Seufert, S., & Szarvas, G. (2018). On challenges in machine learning model management. IEEE Bulletin Tech. Committee Data Engineering.
  • Scherer, M. U. (2015). Regulating artificial intelligence systems: Risks, challenges, competencies, and strategies. Harvard Journal of Law & Technology, 29, 353.
  • Schneider, J., & Handali, J. P. (2019). Personalized explanation for machine learning: A conceptualization. In European Conference on Information Systems (ECIS).
  • Serban, A., van der Blom, K., Hoos, H., & Visser, J. (2020). Adoption and effects of software engineering best practices in Machine Learning. Proceedings of the 14th ACM/IEEE International Symposium on Empirical Software Engineering and Measurement (ESEM), 1–12. 10.1145/3382494.3410681
  • Shirer, M., & Daquila, M. (2019). Worldwide spending on artificial intelligence systems will be nearly 98 billion in 2023. Retrieved November 01, 2020, from https://www.idc.com/getdoc.jsp?containerId=prUS48191221
  • Sidorova, A., & Rafiee, D. (2019). AI agency risks and their mitigation through business process management: A conceptual framework. In Hawaii International Conference on System Sciences (HICSS). 10.24251/hicss.2019.704
  • Siebert, J., Joeckel, L., Heidrich, J., Nakamichi, K., Ohashi, K., Namba, I., Yamamoto, R., Aoyama, M. (2020). Towards guidelines for assessing qualities of machine learning systems. In International Conference on the Quality of Information and Communications Technology (pp. 17–31). 10.1007/978-3-030-58793-2_2
  • Silver, D., Schrittwieser, J., Simonyan, K., Antonoglou, I., Huang, A., Guez, A., Hubert, T., Baker, L., Lai, M., Bolton, A., Chen, Y., Lillicrap, T., Hui, F., Sifre, L., van den Driessche, G., Graepel, T., & Hassabis, D. (2017). Mastering the game of go without human knowledge. Nature, 550(7676), 354–359. doi:10.1038/nature24270
  • Smit, K., & Zoet, M. (2020). A review of AI principles in practice. In Proceedings of the Pacific-Asia Conference on Information Systems (PACIS).
  • song, H., Kim, M., Park, D., & Lee, J.-G. (2020). Learning from noisy labels with deep neural networks: A survey. arXiv, 2007.08199.
  • Sridhar, V., Subramanian, S., Arteaga, D., Sundararaman, S., Roselli, D., & Talagala, N. (2018). Model governance: reducing the anarchy of production ML. In Proceedings of USENIX Annual Technical Conference, 351–358.
  • Stecher, P., Pohl, M., & Turowski, K., 2020. Enterprise architecture’s effects on organizations’ ability to adopt artificial intelligence-A Resource-based perspective. In Proceedings of the European Conference on Information Systems (ECIS).
  • Stoica, I., Song, D., Popa, R. A., Patterson, D., Mahoney, M. W., Katz, R., Joseph, A.D., Jordan, M., Hellerstein, J.M., Gonzalez, J.E., Goldberg, K. (2017). A Berkeley view of systems challenges for AI. arXiv Preprint arXiv, 1712.05855.
  • Tallon, P. P., Ramirez, R. V., & Short, J. E. (2013). The information artifact in IT governance: Toward a theory of information governance. Journal of Management Information Systems, 30(3), 141–178. 10.2753/MIS0742-1222300306
  • Tarafdar, M., Beath, C. M., & Ross, J. W. (2019). Using AI to enhance business operations 60(4), 37-44. Retrieved November 03, 2020, from 10.7551/mitpress/12588.003.0015
  • Tiwana, A., Konsynski, B., & Venkatraman, N. (2013). Special issue: information technology and organizational governance: the IT governance cube. Journal of Management Information Systems, 30(3), 7–12. 10.2753/MIS0742-1222300301
  • Van Horn, J. D., Fierro, L., Kamdar, J., Gordon, J., Stewart, C., Bhattrai, A., … Jain, P. (2018). Democratizing data science through data science training. In Proccedings of Pacific Symposium on Biocomputing (PSB) (pp. 292–303). 10.1142/9789813235533_0027
  • Vartak, M., Subramanyam, H., Lee, W.-E., Viswanathan, S., Husnoo, S., Madden, S., & Zaharia, M. (2016). Modeldb: A system for machine learning model management. In Proceedings of the Workshop on Human-In-the-Loop Data Analytics. 10.1145/2939502.2939516
  • Verbraeken, J., Wolting, M., Katzy, J., Kloppenburg, J., Verbelen, T., & Rellermeyer, J. S. (2020). A survey on distributed machine learning. ACM Computing Surveys (CSUR), 53(2), 1–33.
  • Verma, S., & Rubin, J. (2018). Fairness definitions explained. In IEEE/ACM international workshop on software fairness (fairware) (pp. 1–7). 10.1145/3194770.3194776
  • Wang, H., Li, C., Gu, B., & Min, W. (2019). Does AI-based credit scoring improve financial inclusion? Evidence from online payday lending. In International Conference on Information Systems(ICIS) (pp. 1–9).
  • Wang, Y., Xiong, M., & Olya, H. (2020). Toward an understanding of responsible artificial intelligence practices. In Proceedings of The Hawaii International Conference on System Sciences(HICSS). 10.24251/hicss.2020.610
  • Watanabe, Y., Washizaki, H., Sakamoto, K., Saito, D., Honda, K., Tsuda, N., Fukazawa, Y., Yoshioka, N. (2019). Preliminary systematic literature review of machine learning system development process. arXiv preprint arXiv:1910.05528. 10.1109/compsac51774.2021.00207
  • Webster, J., & Watson, R. T. (2002). Analyzing the past to prepare for the future: Writing a literature review. MIS quarterly, xiii–xxiii.
  • Winfield, A. F., Michael, K., Pitt, J., & Evers, V. (2019). Machine ethics: The design and governance of ethical ai and autonomous systems. Proceedings of the IEEE, 107( 3). 10.1109/jproc.2019.2900622
  • Wirth, R., & Hipp, J. (2000). CRISP-DM: Towards a standard process model for data mining. In Proceedings of the International Conference on the Practical Applications of Knowledge Discovery and Data Mining (pp. 29–39).
  • Witten, I. H., & Frank, E. (2017). Data mining: Practical machine learning tools and techniques (4th edition ed.). Morgan Kaufmann.
  • Zapadka, P., Hanelt, A., Firk, S., & Oehmichen, J., 2020. Leveraging “AI-as-a-Service”–antecedents and consequences of using artificial intelligence boundary resources. In International Conference on Information Systems (ICIS).
  • Zhang, J. M., Harman, M., Ma, L., & Liu, Y. (2020). Machine learning testing: Survey, landscapes and horizons. IEEE Transactions on Software Engineering. 10.1109/tse.2019.2962027