44,327
Views
62
CrossRef citations to date
0
Altmetric
Articles

Ethical issues in focus by the autonomous vehicles industry

ORCID Icon, , ORCID Icon & ORCID Icon
Pages 556-577 | Received 04 Jan 2020, Accepted 25 Nov 2020, Published online: 07 Jan 2021

ABSTRACT

The onset of autonomous driving has provided fertile ground for discussions about ethics in recent years. These discussions are heavily documented in the scientific literature and have mainly revolved around extreme traffic situations depicted as moral dilemmas, i.e. situations in which the autonomous vehicle (AV) is required to make a difficult moral choice. Quite surprisingly, little is known about the ethical issues in focus by the AV industry. General claims have been made about the struggles of companies regarding the ethical issues of AVs but these lack proper substantiation. As private companies are highly influential on the development and acceptance of AV technologies, a meaningful debate about the ethics of AVs should take into account the ethical issues prioritised by industry. In order to assess the awareness and engagement of industry on the ethics of AVs, we inspected the narratives in the official business and technical reports of companies with an AV testing permit in California. The findings of our literature and industry review suggest that: (i) given the plethora of ethical issues addressed in the reports, autonomous driving companies seem to be aware of and engaged in the ethics of autonomous driving technology; (ii) scientific literature and industry reports prioritise safety and cybersecurity; (iii) scientific and industry communities agree that AVs will not eliminate the risk of accidents; (iv) scientific literature on AV technology ethics is dominated by discussions about the trolley problem; (v) moral dilemmas resembling trolley cases are not addressed in industry reports but there are nuanced allusions that unravel underlying concerns about these extreme traffic situations; (vi) autonomous driving companies have different approaches with respect to the authority of remote operators; and (vii) companies seem invested in a lowest liability risk design strategy relying on rules and regulations, expedite investigations, and crash/collision avoidance algorithms.

Introduction

The onset of autonomous driving has provided fertile ground for discussions about ethics in recent years. In addition to the ongoing debates regarding ethical issues particular to automated driving systems-equipped vehicles, the disruptive yet mundane nature of this technology dictated its archetypal role in broader conversations about ethics and Artificial Intelligence (Etzioni & Etzioni, Citation2017; Hengstler, Enkel, & Duelli, Citation2016; Lin, Abney, & Jenkins, Citation2017). The vehicle at the centre of these discussions is a machine designed to provide conveyance on public streets, roads, and highways with conditional, high, or full driving automation (SAE On-Road Automated Vehicle Standards Committee, Citation2014). Such a machine is commonly known as Autonomous Vehicle (AV)Footnote1 and is expected to yield a multitude of social benefits to traffic safety, mobility, and accessibility (Anderson et al., Citation2014; Fleetwood, Citation2017; Milakis, Van Arem, & Van Wee, Citation2017).

The ethics discussions associated with AVs which are documented in the scientific literature have mainly revolved around extreme traffic situations depicted as moral dilemmas, i.e. situations in which the AV is required to make a difficult moral choice between actions in traffic which will result in different combinations of lives saved and sacrificed (Awad et al., Citation2018; Bonnefon, Shariff, & Rahwan, Citation2016; Goodall, Citation2016; Lundgren, Citation2020). Scholars have debated exhaustively the relevance of the AV moral dilemma (Himmelreich, Citation2018; Keeling, Citation2020; Wolkenstein, Citation2018), the merits of using different ethical frameworks, such as Deontology, Utilitarianism, or Rawlsianism, as control algorithms for AVs (Bergmann et al., Citation2018; Huang, Greene, & Bazerman, Citation2019; Keeling, Citation2017; Leben, Citation2017; Lin, Citation2016), and the moral preferences and societal expectations about the ethics to be encoded in AVs (Awad et al., Citation2018; Bonnefon et al., Citation2016). Other ethical issues have been addressed in the scientific literature, such as ethical design, accountability, human meaningful control, sustainability, and privacy (Aydemir & Dalpiaz, Citation2018; Collingwood, Citation2017; Endsley, Citation2017; Hevelke & Nida-Rümelin, Citation2015; Santoni de Sio & Van den Hoven, Citation2018; Taiebat, Brown, Safford, Qu, & Xu, Citation2018).

Quite surprisingly, little is known about the ethical issues in focus by the AV industry. General claims have been made about the struggles of companies over such issues but these lack proper substantiation (Awad et al., Citation2018; Fagnant & Kockelman, Citation2015; Kirkpatrick, Citation2015). As private companies are highly influential on the development and acceptance of AV technologies (Van den Hoven, Vermaas, & Van de Poel, Citation2015), their stance on ethics should be taken into account for the purposes of a meaningful debate about the ethics of AVs.

In order to assess the ethics awareness and engagement of industry, we inspected the narratives in official business and technical reports of companies operating in the AV field. In this research, we focused on the companies with an AV testing permit in California, where there was an early adoption of comprehensive regulations governing the testing of AVs (Favarò, Eurich, & Nader, Citation2018; Soriano, Dougherty, Soublet, & Triepke, Citation2014). We believe that the analysis of reports from a wide range of technology and manufacturing companies in the forefront of AV technology allows us to draw important insights about ethics within the AV industry.

We first provide an overview of the ethics narratives both in the scientific literature and industry reports. At this point, it should be noted that the main aim of this paper is not to present an exhaustive review of the scholarly literature concerning ethical issues surrounding the development and deployment of AVs. Rather, our aim is to explore how the discussion of ethical issues in industry reports and its counterpart in the academic literature relate and compare to one another.

For reasons of brevity, we focus on the matters of safety and cybersecurity, accountability, and human oversight, control, auditing of AVs as presented in the scientific literature thus raising critical yet practical questions for which we will look for answers in the industry narratives. These three issues, which we selected as our focus points, have generated a particularly rich debate in both streams of literature, and are often discussed in relation to one another. We expect that, by providing empirical insights from industry, we can make a contribution for a richer, less speculative, and more meaningful debate on the ethics of AVs.

Methodology

The ambiguous nature of ethics makes systematisation challenging. Here, we attempt to alleviate ambiguity by building our research around a list of ethical issues compiled from 22 major guidelines of AI ethics (Hagendorff, Citation2020). We use this list of AI ethical issues to guide us in identifying the ethics within the scientific and industry narratives.

Interestingly, Hagendorff is reluctant about the effectiveness of AI ethics guidelines. He argues that these sorts of guidelines, traditionally based on a deontological approach to ethics which relies on fixing a set of principles and maxims, should be augmented with a virtue ethics oriented approach aiming at addressing values, attitudes, and behavioural dispositions that would ultimately help professionals refraining from unethical actions (Hagendorff, Citation2020). We acknowledge the limitations of these deontology-based ethics guidelines in promoting a robust ethics culture within organisations. And moreover, it is noted that, because we are using a list of ethical issues based on deontological guidelines as a guidance tool in this research, our results will necessarily reflect such top-down deontological approach, thus leaving out other potential relevant ethical approaches and principles related, for instance, to informed consent and risk acceptance (Menon & Alexander, Citation2020).

While acknowledging these limitations, we believe this list is adequate for our research given that it includes a comprehensive and state-of-the-art compilation of ethical issues in the field of AI ethics.

The original list featuring 22 ethical issues in published guidelines about AI (Hagendorff, Citation2020) was adjusted for this research. We removed one ethical issue (“field-specific deliberations”) as well as all “AI” references as we focus on the AV as a particular AI-powered technology. The final list of 21 ethical issues can be found in .

Figure 1. Ethical issues in AV industry.

Figure 1. Ethical issues in AV industry.

We first reviewed the scientific literature, with the aim of outlining the AV ethics debates, by identifying the ethical issues prioritised by the scientific community and the main empirical findings.

For this purpose, we followed the Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) framework (Liberati et al., Citation2009). Using the combination of keywords “Autonomous Vehicles” AND “Ethics”; “Autonomous Vehicles” AND “Moral”; “Self driving” AND “Ethics”; “Self driving” AND “Moral”; “Driverless” AND “Ethics”; and “Driverless” AND “Moral” in Scopus and Google Scholar (in the latter only records within the 2015–2020 timeframe in the first five pages of the database were considered), 715 records were identified. Five additional records were identified through other sources, such as citation chaining. Upon initial screening of the 720 records, 324 duplicates were removed, which meant that 396 records were assessed for eligibility. Only published scientific documents with available full text written in English language and addressing the moral or ethics dimensions of automated driving systems were considered eligible for this study. Therefore, based on this eligibility criteria, 158 records were excluded (23 records were not published scientific documents; 47 records did not have full text available; 3 records were not written in English language; and 85 records were not about the moral or ethics dimensions of automated driving systems) thus leaving a total of 238 records for further review and analysis.

It is recalled that, as far as this literature review is concerned, our aim is quite modest as we intend solely to provide an outline of the AV ethics debates in the scientific literature. We started by reading and analysing the articles in order to identify the ethical issues prioritised by the scientific community. Thereafter, we divided the articles into theoretical (195) or empirical (43), depending on the type of research employed. And finally we reflected on the theoretical propositions and main empirical findings related to the ethical issues which are the focus points of this research.

Following the review of the scientific literature, we proceeded to the document review of AV business and technical reports, relying on the list of ethical issues mentioned above, to identify the relevant issues within the industry narratives. For the selection of companies, we used the record of 66 companies with an AV testing permit both with and without a driver in California made available by the Department of Motor Vehicles as of June 2020.

The technical and business reports from the past five years were requested from the companies and also screened through standard online searches on their websites. For reasons of reproducibility of this research we only considered reports that could be downloaded and saved as portable document format (pdf) files thus excluding articles, blog entries, or other materials made available by the companies online but which carried the risk of not being accessible in the future. In total, we used 86 documents from 29 companies.

These reports are curated documents that serve the purpose of communicating corporate information to investors, consumers, and regulatory agencies. For that reason, such reports may depict augmented or abbreviated accounts of the range of actions taken by AV companies with respect to ethics. These documents are, nevertheless, important pieces of information to learn the industry's formal stance on the complex ethical issues associated with AVs.

Initially, we proceeded with the reading of the documents and selection of statements that signalled ethical considerations in the context of AVs. Although contextual analysis is crucial for this investigation, we acknowledge the limitations of the manual approach. Therefore, on a second occasion, we relied on linguistic-based text data analytics in order to assess the validity of our initial results. We started by creating lexicons, i.e. groups of search keywords organised to investigate a concept (Schuelke-Leech, Jordan, & Barry, Citation2019), associated with each ethical issue. Subsequently, we applied a text mining algorithm using the previously created lexicons as regular expressions in order to locate the keywords associated with each ethical issue in the 86 documents. The output generated by this algorithm is a report stating the number of occurrences of the keywords in the lexicons associated with each ethical issue in each one of the documents issued by the AV companies. And lastly, we compared the results of the text mining algorithm and the manual approach and made the necessary adjustments with reference to the contexts of the narratives. For the quantitative analysis of the results, we did a standard descriptive statistical analysis of the ethical issues found in the AV industry reportsFootnote2.

Overview of the ethics of AV technology in scientific literature

The amount of attention that AV technology ethics has received in recent years is quite new to the field of Transportation. Traditionally, the ethics debates in this field have revolved around less sensational issues, such as cost–benefit analysis of transport projects or fairness in pricing (Van Wee, Citation2011). The advent of autonomous driving is a remarkable scientific and engineering achievement that has given rise to novel and controversial ethical issues.

Our review showed quite clearly that the scientific literature on AV ethics is dominated by considerations about safety and cybersecurity concerning the programming of extreme traffic situations. This controversial issue is commonly known as the trolley problem in reference to a thought experiment popularised by Philippa Foot in 1967 in which an agent needs to make a difficult choice of allowing a runway trolley to proceed its course and kill five track workers or divert the trolley from its course killing only one worker (Foot, Citation1967; Thomson, Citation1984). There are many variations and extensions to this thought experiment but its core can be defined as a moral choice between actions in traffic which will result in different combinations of lives saved and sacrificed. Because extreme traffic situations need to be programmed in advance, AV technology seemed to bring this textbook thought experiment to life thus capturing the attention of scholars and the media.

We found references to the trolley problem in more than half of the 238 reviewed articles. Most of these articles are theoretical pieces of research, often written as argumentative or normative essays, about different perspectives and dimensions of the AV moral dilemma. While these debates are certainly very rich, we found this stream of the literature to be quite fragmented. For instance, there is still little consensus about the relevance of the trolley problem in the context of AVs (Bonnefon, Shariff, & Rahwan, Citation2019; Cunneen et al., Citation2020; Himmelreich, Citation2018; Keeling, Citation2020; Wolkenstein, Citation2018).

The empirical findings reported in the literature are also quite controversial, as they reveal potential challenges in adapting societal expectations to moral decision-making driving algorithms (Kallioinen et al., Citation2019). The AV social dilemma, i.e. a conflict between individual and collective interest in the context of autonomous driving technology, illustrates such challenge. It has been reported that people approve and would like others to buy utilitarian AVs which sacrifice their passengers for the greater good, yet prefer to ride in AVs that protect their passengers at all costs thus disapproving utilitarian regulation of AVs (Bonnefon et al., Citation2016; Huang et al., Citation2019; Morita & Managi, Citation2020).

A substantial amount of research, namely the Moral Machine Experiment (MME) (Awad et al., Citation2018), has focused on collecting and analysing moral preferences and societal expectations about the ethics to be encoded in AVs. However, the methodological soundness and value of such investigations for the purpose of defining moral algorithms for AVs have been questioned (Bigman & Gray, Citation2020; Harris, Citation2020). The main contribution of the MME, regarding moral preferences in AV moral dilemmas, conflicts with current ethical guidelines, such as rule 9 of the German Ethics Code for Automated and Connected Driving, which prohibits distinctions based on personal features in the case of unavoidable accident situations (Harris, Citation2020; Kallioinen et al., Citation2019; Luetge, Citation2017). The proponents of the MME acknowledge that AV policy should not necessarily follow public expectations and preferences but they believe that such preferences should not be completely dismissed. They argue that, given the strong preference for sparing children, it would be challenging to explain the rationale for not assigning a special status to children (Awad et al., Citation2018). Recently, it has been hypothesised that AV fatalities carry more weight because those are rare events and not so much due to intrinsic differences in public perception between AV and conventional vehicles fatalities (Huang, van Cranenburgh, & Chorus, Citation2020).

Several scholars have reported an overstatement of the AV trolley problem and called for the ethics community to focus on other ethical issues associated with AVs (Goodall, Citation2019; Himmelreich, Citation2018; Keeling, Evans, Thornton, Mecacci, & de Sio, Citation2019).

Other issues debated in the literature include ethical design, accountability, human meaningful control, sustainability, and privacy (Aydemir & Dalpiaz, Citation2018; Collingwood, Citation2017; Endsley, Citation2017; Hevelke & Nida-Rümelin, Citation2015; Santoni de Sio & Van den Hoven, Citation2018; Taiebat et al., Citation2018). Particularly, matters related to accountability and human meaningful control have received considerable attention in the literature recently and, along with safety and cybersecurity, will be further explored later in this research.

Overview of the ethics of AV technology in AV industry reports

In this research, we focus on the AV industry in California, a State that has been an early and strong proponent of this technology and hosts many R&D programs (Brown et al., Citation2018). As of June 2020, the California DMV had listed 66 permit holders for testing with a driver, 2 permit holders for driverless testing, and 0 permit holders for AV deployment.

For our analysis, we used a total of 86 documents issued by 29 companiesFootnote3 in the forefront of AV technology. Therefore, we consider their official reports as important pieces of information about the industry's formal stance on the complex ethical issues associated with AVs.

Overall, the AV reports reviewed in this research show a overwhelmingly positive tone about AV technology, which needs to be interpreted in light of such reports being written for a particular audience of investors, consumers, and regulatory agencies. Although lacking the rigour and depth of the narratives in the scientific literature, a plethora of ethical issues are referenced in the AV industry reports.

A quantitative synopsis of the ethics prioritised by companies with an AV testing permit in California, based on the number of companies that addressed each one of the 21 ethical issues in their reports, can be found in . It shows that safety and cybersecurity is the ethical issue that was addressed by more companies, followed by common good, sustainability, well-being; human oversight, control, auditing; and science-policy link. In contrast, issues such as hidden costs, fairness, non-discrimination, justice, or responsible research funding were not addressed in these reports.

The relevance of particular ethical issues within industry narratives requires a balance between frequency and comprehensiveness. We recall that three ethical issues had previously been selected from the scientific literature to be further explored in this research (safety and cybersecurity, accountability, and human oversight, control, auditing). Safety and human oversight issues are frequently addressed by AV companies and, whereas accountability is addressed by a lower number of companies, it is comprehensively explored in the reports we analysed.

It should be noted that, even when we found quite comprehensive accounts on particular ethical issues, the narratives in the AV reports are consistently pragmatic and oriented towards technical solutions. For instance, in its reports, Mercedes-Benz acknowledges the importance of data privacy while emphasising practical data-protection-friendly solutions that provide privacy by design in compliance with privacy laws (Daimler Sustainability Report 2018 and Reinventing Safety: A Joint Approach to Automated Driving Systems).

Safety and cybersecurity

Safety and security are both related to the integrity of systems but, whereas safety concerns the adequate functioning of a system, security is about the ability of a system to resist intentionally malicious actions. There are unsettled considerations about the acceptable safety and cybersecurity levels of AVs, both in mundane and extreme situations, in order to secure the well-being of users and other traffic agents (Kalra & Paddock, Citation2016; Parkinson, Ward, Wilson, & Miller, Citation2017; Sparrow & Howard, Citation2017).

Safety and cybersecurity of AVs in the scientific literature

Mundane and extreme traffic situations

Mundane traffic situations are the day-to-day interactions of traffic agents (pedestrians, cyclists, animals) that require some flexibility, such as crossroads, highway entrances, or crosswalks with limited visibility. These interactions are challenging for AVs not only because these systems lack human intuition and flexibility but also because of the large scale fleet programming that is needed (Borenstein, Herkert, & Miller, Citation2019; Himmelreich, Citation2018). Coordination over different technical approaches to these traffic situations is crucial to ensure safety but it is unclear how such technological coordination can be facilitated in the competitive AV market (Himmelreich, Citation2018).

Extreme traffic situations are unexpected occurrences in the traffic environment which entail danger for vehicle occupants and other traffic agents (e.g. the unexpected appearance of an animal on a highway). Some of these situations are depicted in the scientific literature as moral dilemmas. As mentioned above, these difficult moral situations are highly explored in the scientific literature with reference to the trolley problem thought experiment (Foot, Citation1967; Thomson, Citation1984).

The underlying argument in the debates that take place in the scientific literature about the AV trolley problem can be outlined as follows: (i) AVs ought to save lives. (ii) However, upon deployment of AVs, extreme traffic situations will not be completely avoided. (iii) Some of these extreme traffic situations will require AVs to make difficult moral decisions. (iv) Difficult moral decisions in traffic resemble the trolley problem. (v) The best option to assist AVs in managing the AV trolley problem is x. (vi) Option x is programmable. (vii) Therefore AVs should be programmed with x. The disputes in the literature about the AV trolley problem are mainly related to premises (iv) regarding the relevance of the trolley problem in the AV context and (v) regarding the merits of different approaches to assist AVs in moral decision-making in extreme traffic situations.

Relevance of the trolley problem

The relevance of the trolley problem in the AV context concerns its value as a model to investigate a relevant AV ethical challenge. It has been acknowledged that using trolley cases as inputs for crash optimisation algorithms invites a myriad of criticism (Keeling, Citation2020; Nyholm & Smids, Citation2016). Scholars have argued that trolley cases are of limited usefulness for the ethics of AVs because such cases would not only be highly improbable occurrences, but also their assumptions are unrealistic (outcomes of the different moral decisions available to the agent are known rather than probabilistic), inconsistent (agent has control over a vehicle yet a collision is imminent and unavoidable), and limited with respect to design (trolley cases assume a top-down approach in which an agent makes a decision explicitly, thus failing to encompass different design approaches to decision-making) (Fried, Citation2012; Himmelreich, Citation2018).

It is accepted that trolley cases are dramatic, stylised, black-and-white situations that have little resemblance to real-life extreme traffic situations. However, it is also widely acknowledged that AVs will not eliminate crashes (Bagloee, Tavana, Asadi, & Oliver, Citation2016; Bonnefon et al., Citation2019; Favarò, Nader, Eurich, Tripp, & Varadaraju, Citation2017; Fleetwood, Citation2017). Therefore, weak trolley cases seem to be plausible. An example of a weak trolley case from the literature is an AV which is travelling across a two-lane bridge when a bus in the other lane swerves into its lane and the AV needs to decide either to brake, which would result in a collision with the bus, or to swerve into the other lane, thus hitting the side of the bridge (Goodall, Citation2014a). Such extreme traffic situations, entailing decisions about who is put at marginally more risk of being sacrificed, may be rare occurrences when AVs are deployed but they need to be addressed (Bonnefon et al., Citation2019).

Recently, it has been proposed that the relevance of the trolley cases in the AV context is associated with the prospect of development of novel ethical principles. These principles, formulated upon analyses of the moral intuitions that emerge in stylised cases, would ultimately guide the AV design process (Keeling, Citation2020; Wu, Citation2019).

Approaches to assist AVs in extreme traffic situations

Another debate in the scientific literature concerns the conflicting approaches that have been advanced by scholars to address extreme traffic situations. We recall premise (v) above, The best option to assist AVs in managing such extreme traffic situations is x, to clarify that in the AV ethics literature x tends to be proposed within the realm of Machine Ethics.

Scholars have debated the merits of using ethical frameworks such as Rawlsianism, Deontology, or Utilitarianism as the control algorithms of AVs (Bergmann et al., Citation2018; Huang et al., Citation2019; Keeling, Citation2017; Leben, Citation2017; Lin, Citation2016; Thornton, Pan, Erlien, & Gerdes, Citation2017) or aggregating societal moral preferences (Awad et al., Citation2018; Etienne, Citation2020; Harris, Citation2020; Noothigattu et al., Citation2018) to encode ethics in AVs, thus assisting them in navigating extreme traffic situations that would require moral choices.

It was shown earlier that safety considerations are central both in scientific literature and industry reports. However, scholars have mostly debated the ethics in extreme traffic situations with reference to trolley cases. We further investigate the industry's approach to extreme traffic situations and raise two relevant questions: (i) Are extreme traffic situations resembling trolley cases addressed by industry? and (ii) What are the solutions proposed by industry to address extreme traffic situations?

Safety and cybersecurity in the AV industry reports

Safety and trust

Considering that the commercial success of AV technology depends greatly on the trust of consumers, it is hardly surprising that the industry narratives focus mainly on safety issues. In A Matter of Trust Ford's Approach to Developing Self-Driving Vehicles, it is stated that for autonomous vehicles to be accepted by the public it needs to be established that they can be trusted (Marakby, Citation2018) and in Intel's white paper A Matter of Trust: How Smart Design Can Accelerate Automated Vehicle Adoption, trust is also emphasised when it is stated that before driverless AVs can be widely accepted, people must be willing to trust them with their lives and the lives of those they care about hence AVs must behave, react, and communicate in ways that make it easy for people to trust them – not only the passengers inside, but also pedestrians and the other drivers who encounter them on the road (Weast, Yurdana, & Jordan, Citation2016).

Trust and business-related considerations may not be the only reasons for the prevalence of safety considerations in the AV reports. Autonomous driving is a complex and disruptive technology which is expected to have a major societal impact. Unlike other social and ethical issues, such as fairness or human autonomy in an AI-dominated society, safety challenges are prone to be solved by technical or engineering approaches (Hagendorff, Citation2020). Therefore, AV companies tend to prioritise these issues, for which technical solutions are presented.

In the reports reviewed in this research, we found extensive safety considerations both for mundane and extreme traffic situations. In order for AVs to successfully deal with mundane traffic situations, companies propose advanced sensing and AI-powered solutions. Mercedes-Benz and Bosch designed a Object and Event Detection and Response (OEDR) system for AVs which is based on sensors, actuators, and computing resources that is expected to assist the AV in handling these traffic situations - Reinventing Safety: A Joint Approach to Automated Driving Systems (Daimler, Citation2018), whereas Valeo proposes an AI-based approach, building on the thought that in order to negotiate complex traffic conditions where there are many unknowns, AVs need to learn the data – Meet the Future 2016 Activity and Sustainable Development Report (Valeo, Citation2016).

Extreme traffic situations: crashworthiness, collisions, and moral dilemmas

Safety considerations with respect to extreme traffic situations are also explored in the industry reports. Companies focus on the crashworthiness of AV technology, which is quite relevant for our investigation of the AV moral dilemma as, at its core, the trolley case – either in its weak or strong version – is a convoluted crash optimisation problem. By inspecting the industry reports regarding crashworthiness, we expect to clarify some of the critical elements of the AV moral dilemma.

The first element concerns the risk of crashing. Indeed the AV moral dilemma could be promptly dismissed on the account that autonomous driving will eliminate crashes. While companies express their vision of a future without accidents (Advanced Driver Assistance Systems continues to evolve in order to realise autonomous driving and zero-accident smart vehicles, the essence of the fourth industrial revolutionMando Sustainability Report Mando, Citation2018), such ambition is mitigated by the plethora of statements on the inevitability of AV crashes and collisions which leave no room to entertain the thought of a complete elimination of accidents (driving environments can be extremely complex and difficult and no automated driving system – regardless of how capable it may be – is likely to prevent crashes entirelyAutomated Driving at Toyota: Vision, Strategy and Development (Toyota, Citationn.d.); While our top priority is to avoid collisions, we recognise it is possible that we could be involved in a collision at some pointDelivering Safety: Nuro's Approach Nuro, Citation2018).

Accepting that the future will not be crash and collision-free leads us to further considerations about the AV moral dilemma. Indeed crashes and collisions are a necessary condition for such extreme situations. In our document review, we did not find any reference to trolley cases as described in the scientific literature, i.e. situations that require the AV to make difficult moral choices (Bonnefon et al., Citation2016; Goodall, Citation2014b), but we identified nuanced allusions to this matter.

Companies acknowledge that AVs will face rare extreme traffic situations, often mentioned in the industry reports as edge cases, and emphasise simulation and validation methods used to test these scenarios (we test and validate our self-driving vehicles in the wide variety of environmental conditions that the vehicle might face in its operational design domain – from driving scenarios the vehicle would face daily to the rare edge casesGeneral Motors Self-Driving Safety Report (GM, Citation2018); AI-powered autonomous vehicles must be able to respond properly to the incredibly diverse situations they could experience, such as emergency vehicles, pedestrians, animals, and a virtually infinite number of other obstacles – including scenarios that are too dangerous to test in the real worldNvidia Self-Driving Safety Report (Nvidia, Citation2018); decision making is one of the most challenging tasks in the A.I. development of an autonomous vehicle … there are infinite edge cases that may be difficult or dangerous to reproduce in reality, such as illegal driving behaviours or sudden traffic accidentsThe Autox Safety Factor AutoX, Citation2018).

We found one statement that somewhat resembles the AV moral dilemma, with one important caveat regarding the nature of harms at stake. In the Nuro report Delivering Safety: Nuro's Approach, it is stated that, in the unlikely case of a Nuro shuttle ever encountering an unavoidable collision scenario the driverless passengerless vehicle has the unique opportunity to prioritise the safety of humans, other road users, and occupied vehicles over its contents (Nuro, Citation2018). Whereas we can not legitimately consider Nuro's account as an AV moral dilemma, we consider it as yet another indication that companies are aware of convoluted situations akin to weak versions of moral dilemmas. We speculate that Nuro's slightly more transparent stance on this matter could be explained by the fact that it focuses on passengerless self-driving delivery technology (Our custom vehicle is engineered to make delivery of everything more accessible – from groceries to pet food, prescription drugs to dry cleaning … with no driver or passengers to worry about, our vehicle can be built to keep what's outside even safer than what's inside … it's lighter, nimbler, and slower than a passenger car, and is equipped with state-of-the-art software and sensing capabilities that never get distracted Nuro, Citation2018).

At the root of extreme or edge cases are often blind spots that prevent the AV from performing an accurate evaluation of the traffic context and having enough emergency braking time. The solutions advanced by companies to address this problem rely on radars and speed limitation when the visual field of the AV is obstructed. Although blind spot detection and assistance is considered a low level automation feature, the narratives we found about this issue and its implications for pedestrians' safety are yet another substantiation of the concerns of AV companies about extreme traffic situations (in the case of pedestrians who are occluded from the vehicle … it should adjust the speed such that if a child would emerge from behind some object there would be no accident … even in a worst case scenario where the pedestrian emerges from behind some sensing obstruction (e.g. a parked car) even at that maximal speed - Intel Implementing the RSS Model on NHTSA Pre-Crash ScenariosFootnote4 (Mobileye, Citationn.d.); If the view is blocked Perception will flag that area as unknown … if an object is hard to see because of rain or fog or because it is hidden behind a truck the computer brain knows that and adjusts its decision-making and performance accordingly … this allows prudent decision-making and operation based upon both what the sensors “see” as well as what may be hidden from view GM, Citation2018).

As a result of our review of AV reports, we conclude that moral dilemmas resembling trolley cases are not addressed in these reports in the terms described in scientific and media publications, but there are nuanced allusions that unravel underlying concerns about these extreme traffic situations. Regarding the solutions proposed by industry to address extreme traffic situations, we report for now that companies rely on radars and speed limitation to address the problem of blind spots which are often at the root of these traffic situations.

Human oversight, control, auditing of AVs

Human oversight, control, and auditing of autonomous systems imply the surveillance of the development and performance of the technology. It is expected that remote oversight of the performance of autonomous driving ensures trust and safety in this technology as human operators are able to take-over the vehicle. It has been reported, however, that several technical problems take place precisely during the transfer of control over the AV (Heikoop et al., Citation2019).

Human oversight, control, auditing of AVs in the scientific literature

A philosophical account of meaningful human control over automated systems has been proposed by Santoni de Sio and Van den Hoven to ground the design guidelines with respect to human oversight, control, and auditing of AV technology (Fischer & Ravizza, Citation2000; Santoni de Sio & Van den Hoven, Citation2018). According to this account, AVs should meet tracking and tracing conditions to allow for a meaningful form of human control (Santoni de Sio & Van den Hoven, Citation2018). An AV should therefore be able to track the relevant human moral reasons in a sufficient number of occasions, thus adjusting its behaviour in accordance to the intentions of a human designer or operator – tracking condition – and its actions should be traceable to a proper moral understanding on the part of the humans who design and deploy the system – tracing condition (Santoni de Sio & Van den Hoven, Citation2018).

The tracking condition ensures that the AV complies with the intentions of a human operator. It should be noted, however, that humans are poor fallback systems. As more autonomy is added to a system and it becomes more reliable and robust, the situation awareness of human operators decreases and they are less likely to take over manual control (automation conundrum) (Endsley, Citation2017). Therefore, in critical situations, it could be that, by meeting the tracking condition, the AV is complying with an instruction issued by a low situation awareness operator.

The tracing condition requires the presence of at least one human agent that can understand the real capabilities of the system and bear the moral consequences of the actions of the system (Santoni de Sio & Van den Hoven, Citation2018). This condition is especially relevant to tackle the responsibility gaps, i.e. situations where it is unclear who should be responsible for an outcome (Matthias, Citation2004; Nyholm, Citation2018), that are expected to arise in the context of AV technology as a result of the fragmentation of the technology action (many-hands problem Van de Poel, Fahlquist, Doorn, Zwart, & Royakkers, Citation2012).

Meaningful human control has been heralded as the standard for AVs to meet the appropriate level of safety and accountability (Keeling et al., Citation2019). We will revisit this theory in the section below about the accountability of AVs. From this section, we raise one relevant question, regarding the tracking condition, to be investigated in the industry reports: according to the autonomous driving industry, which decision prevails in traffic, the decision of the AV or the decision of the human operator?

Human oversight, control, auditing in the AV industry reports

Remote and onsite human oversight of AV operations is addressed in the industry reports analysed in this research. We recall that few companies have a driverless testing permit in California, which means that in general companies rely heavily on onsite human oversight for the testing of AVs. Mission Specialists are trained on the governing operational design domain, and are prepared to take manual control of the vehicle when presented with a scenario that is not included in the current operational design domainUber Advanced Technologies Group A Principled Approach To Safety (Uber, Citation2018). In addition to onsite oversight, companies also rely on the remote control of AV operations. In a report issued by Zoox, it is stated that their remote operations support centre will have operators available to remotely guide vehicles at any time, day or night, when a vehicle encounters an uncertain driving situation such as a traffic light outage or a road obstructionSafety Innovation at Zoox: Setting the bar for safety in autonomous mobility (Zoox, Citation2018).

Building on the notion of tracking, introduced above in the context of Meaningful Human Control, we report different approaches with respect to the authority of remote operators. Companies such as Mercedes-Benz and Intel seem to prioritise the autonomy of the vehicle (while automated driving vehicles take under consideration data received from an infrastructure, particularly data that can be strongly authenticated and validated, the vehicles ultimately maintain their own decision authority, not the infrastructureSafety First of Automated Driving 2019 Wood et al., Citation2019) whereas other companies, such as AutoX, seem to prioritise the decisions made by remote operators (operators at the remote support system can check the AI decision results and correct or overwrite them when unexpected errors occur AutoX, Citation2018).

The statements that we identified in the AV reports regarding human oversight relate more to the first condition of the Human Meaningful Control theory, but we also report one statement which relates to the tracing condition with respect to the understanding of the system. AImotive states that test operators face their own unique challenges. The debug screen of a complex autonomous system is incomprehensible to the untrained eye. These engineers and developers have a deep understanding of the code at work in our prototypes allowing them, at times, to predict when the system may fail. This allows our test crews to retake control of the vehicle preemptively, in a controlled mannerEnsuring Safe Self-Driving AImotive's Development Puts Safety First (Csizmadia, Citation2018). The tracing condition in the Meaningful Human Control theory has another dimension, related to responsibility, which will be addressed below.

Accountability

Accountability issues associated with AV technology have received substantial attention in the scientific literature. We refer to accountability in broad terms, thus encompassing closely related concepts, such as responsibility and liability. It is clarified that accountability entails responsibility, but unlike the latter it requires explanations about actions and it cannot be shared (Mulgan, Citation2000); responsibility for an action traditionally requires at least a control condition, i.e. an agent is responsible if it is the agent of the action, and an epistemic condition, i.e. awareness or knowledge of the agent regarding the action (Coeckelbergh, Citation2020); and liability is legal or financial responsibility (Collingwood, Citation2017). These matters are challenging in the AV domain, mainly because of the fragmentation of the technology action, which can result in responsibility gaps.

Accountability in the scientific literature

Different approaches to AV responsibility have been proposed in the literature (Borenstein, Herkert, & Miller, Citation2017; Coeckelbergh, Citation2016; Misselhorn, Citation2015; Nyholm, Citation2018). The theory of Human Meaningful Control, which was introduced above, encompasses a tracing condition that requires the presence of at least one human agent who can bear the moral consequences of the actions of the AV (Santoni de Sio & Van den Hoven, Citation2018). It has been asserted that, in order for the tracing condition to be met in higher-order levels of automation, a transition of responsibility from the driver to designers or remote operators is required. At such levels of automation, how the AV is designed to execute its tasks is more important than how the human driver ought to execute its tasks (Heikoop et al., Citation2019).

An argument has been presented in the scientific literature particularly concerning liability, in which it is roughly stated that AVs have the potential to save lives but crushing liability may discourage manufacturers from developing and deploying AVs, and as such this technology would not meet its potential to save lives (Hevelke & Nida-Rümelin, Citation2015; Marchant & Lindor, Citation2012). As legal scholars are working on extensions to criminal and civil law (Funkhouser, Citation2013; Gurney, Citation2013, Citation2015), it is questioned whether liability legal frameworks should be designed in such a way that would not impede, but rather promote, the development and improvement of AVs (Hevelke & Nida-Rümelin, Citation2015). This argument has been undermined by some scholars who claim that increased manufacturer liability will not be problematic, as AVs will be safer and will bring down the overall cost of litigation and insurance (Garza, Citation2011). Indeed, it has been reported that thus far governments have avoided strict measures in order to promote AV developments (Taeihagh & Lim, Citation2019).

Another issue that has been presented in the literature is the liability dilemma of the AV manufacturer which showcases the conflict between ethics and law when it comes to liability. When designing a crash collision algorithm, a manufacturer is assumed to face three options while balancing ethics and liability: (i) program an algorithm to swerve in a direction that would sacrifice fewer lives but would entail high liability due to compensatory and punitive damages for intentional conduct caused by targeting the sacrificed people; (ii) allow the AV to run its course which would entail a larger number of lives sacrificed but lower liability which would then be restricted to compensatory damages; and (iii) avoid a collision, which if successful, entails that no lives are sacrificed but if unsuccessful, entails the largest number of lives sacrificed, but in either case entails the lowest liability (Wu, Citation2019). It is therefore concluded that what is easier in a lawsuit may not be the more ethical solution (Wu, Citation2019). By featuring a trolley case in the background and making simplistic and general assumptions about the law, the AV liability dilemma suffers from the same sort of shortcomings that have been pointed out earlier about the AV moral dilemma. Despite its limitations, the liability dilemma of the AV manufacturer sheds light on the tension between ethics and liability, which should not be ignored.

The three issues we explored above regarding the transition of responsibility in higher order levels of automation, the liability and technology development argument, and the liability dilemma can be further investigated within the industry narratives by considering the design strategies with respect to accountability. From this section we raise the question: which accountability design strategy is being adopted by the AV industry?

Accountability in the AV industry reports: the case of the super-humanly fast runner

We found several statements in the industry reports that allow us to further reflect on the approach of the industry with respect to accountability. In general, AV companies seem invested in a lowest liability risk design strategy relying on rules and regulations, expedite investigations, and crash/collision avoidance algorithms.

Companies have stated the need for clear rules to be set in advance (It is necessary to put legal frameworks in place in order to clarify where the responsibility lies in case of the occurrence of an accident after the realisation of fully automated drivingHonda Sustainability Report 2015 Honda, Citation2015), while also presenting technical solutions aimed at expediting investigations, such as the use of a “black-box” akin to a flight recorder designed to store accident data, or responsibility algorithms based on mathematical models (With regard to liability issues a “black-box” is required that stores certain data necessary to sort out accident liability which can contribute towards allocating responsibility between human and machine when accidents happenBMW Sustainable Value Report 2016 (BMW, Citation2016); What will happen when a collision occurs? There will be an investigation, which could take months … our solution is to set clear rules for fault in advance, based on a mathematical model … the investigation can be very short and based on facts, and responsibility can be determined conclusively. This will bolster public confidence in AVs when such incidents inevitably occur and clarify liability risks for consumers and the automotive and insurance industriesIntel A Plan to Develop Autonomous Vehicles. And Prove it. Shashua & Shalev-Shwartz, Citation2017).

It was mentioned earlier that companies acknowledge that AV technology will not eliminate accidents, yet some companies are invested in developing AVs which will never cause or be responsible for accidents (By formally defining the parameters of the dangerous situation and proper response, we can say that responsibility is assigned to the party who did not comply with the proper response. Therefore, the Responsibility-Sensitive-Safety model guarantees that when applying it to any “driving policy” (the decision-making mechanism of the AV), the self-driving car will never initiate a dangerous situation and thus, it will never cause an accident (Mobileye, Citationn.d.); Over time, though, Guardian capability will grow steadily as technology improves, with a goal of creating a vehicle never responsible for a crash regardless of errors made by a human driver Toyota, Citationn.d.). In one of the Intel reports (Intel A Plan to Develop Autonomous Vehicles. And Prove it.) it is stated that their Responsibility-Sensitive-Safety system will always brake in time to avoid a collision with a pedestrian unless the pedestrian is running super-humanly fast (Shashua & Shalev-Shwartz, Citation2017). By providing the super-humanly fast runner illustration, Intel is not only emphasising that their AV will not be responsible for a collision with a pedestrian, but it is also promoting trust in their technology.

Notwithstanding the positive accounts found in the industry reports regarding the development of minimally responsible AV technology, we found a statement in Nissan's Financial Information 2018 bracing the company for potential liability losses related to AVs. It is stated that If the autonomous driving technology is developed and its use becomes quickly widespread in the future, the responsibility of automobile manufacturers might be brought into question in connection with the decline in drivers engaged in driving … If the recalls that the Group has implemented for the benefit of customers' safety become significant in volume and amount, the Group would not only incur significant additional expenses but also experience damage to its brand image, which could adversely affect its financial position and business performance (Nissan, Citation2017).

Conclusion

Despite the wealth of discussions about the ethics of AVs, little is known about the awareness and engagement of the industry on this matter. In this research we have provided an overview of the narratives on the ethics of AVs as presented both in scientific literature and in industry reports issued by companies with an AV testing permit in California. Subsequently, we focused on safety, accountability, and human oversight, and we raised critical yet practical questions, for which we looked for answers in the industry narratives. A combination of contextual analysis and text mining techniques was employed to select statements signalling AV-related ethical considerations within the industry reports.

The overall conclusion that can be drawn from our analyses is that industry and academia look at the ethics of AV technology through rather different lenses. For example, while the scientific literature has been largely preoccupied with deep considerations of abstract moral dilemmas (trolley problem), industry reports adopt a much more pragmatic, technology-infused and perhaps overly optimistic narrative when discussing the potential of so-called edge cases where accidents cannot be avoided and loss of life and damage need to be minimised. While this discrepancy may perhaps not come as a surprise to many, it is disappointing to see that on matters that are of such great importance to the general public, science and industry seem to diverge so profoundly. While we certainly do not advise to try and establish some form of agreement between industry's views regarding the ethical issues surrounding AVs and those of academia (which would be a tall order anyway, given the wide variety of such views within industry and within academia), we do believe that it would be valuable to both sides of the aisle to inform one another of one's viewpoints.

More specifically, the findings in this research suggest that: (i) given the plethora of ethical issues addressed in the reports, autonomous driving companies seem to be aware of and engaged in the ethics of autonomous driving technology; (ii) scientific literature and industry reports prioritise safety and cybersecurity; (iii) scientific and industry communities agree that AVs will not eliminate the risk of accidents; (iv) scientific literature on AV technology ethics is dominated by discussions about the trolley problem; (v) moral dilemmas resembling trolley cases are not addressed in industry reports but there are nuanced allusions that unravel underlying concerns about these extreme traffic situations; (vi) autonomous driving companies have different approaches with respect to the authority of remote operators; and (vii) companies seem invested in a lowest liability risk design strategy relying on rules and regulations, expedite investigations, and crash/collision avoidance algorithms.

Clearly, our study has its limitations, which we would like to highlight once more at this point. Despite our efforts to alleviate ambiguity surrounding terms such as accountability, we could not successfully remove it entirely from this research. This is unavoidable as academic ethicists amongst themselves have rather diverging views on what a term like accountability means, and how it should be distinguished from related concepts such as responsibility. It should therefore not come as a surprise that this ambiguity at an abstract level may translate into different (implicit) meanings attached to the same word, in different industry reports. As a consequence, our analysis of these reports which uses a combination of “manual reading” and text mining, risks conflating different meanings attached to the same vocabulary. One promising way to alleviate or at least diminish this problem is to use techniques that are popular in the field of Anthropology, such as participant observation, in-depth interviews and focus groups. These techniques offer a potential window into how particular terminology is being used in the AV industry, as such providing a base for more carefully discussing how different industry actors differ from one another in terms of their approach to, e.g. accountability in the context of AVs and how industry as a whole differs from academia in this regard.

Such techniques could also help remedy a second limitation of our study, which is that we focused on curated reports that were made publicly available by industry actors for a particular audience. Although, as we argued above, we believe that such documents hold important clues regarding the views of industry actors – e.g. in providing insight into how they like to be seen by others – there is clearly scope and need for more and other types of data collection here. For example, participant observation in which a scholar would be allowed to be embedded in an AV-company for a longer period of time, and to do a range of in-depth interviews with employees at various levels of the organisation, is likely to add significantly to our knowledge of industry's dealings with the ethical conundrums that surround the development and deployment of AVs. We trust that our study would provide a useful stepping stone for such follow up research.

Acknowledgments

The authors acknowledge Nicolas Cointe for his assistance in the linguistic-based text data analytics, the three anonymous reviewers for providing helpful comments on earlier versions of the manuscript, and the European Research Council for financial support of this research (ERC Consolidator grant BEHAVE – 724431).

Disclosure statement

No potential conflict of interest was reported by the author(s).

Additional information

Funding

This research has received funding from the European Research Council: Consolidator Grant BEHAVE (grant agreement 724431).

Notes

1 Different nomenclatures are used for highly automated vehicles such as autonomous vehicles, automated vehicles, self-driving cars, or driverless cars (Gandia et al., Citation2019). Here we adopt autonomous vehicles when referring to automated driving systems-equipped vehicles (levels 3, 4, or 5 driving automation systems according to the Society of Automotive Engineers International Taxonomy and Definitions for Terms Related to Driving Automation Systems for On-Road Motor Vehicles Committee, Citation2014) for reasons of consistency with the nomenclature favoured by the industry. In this context autonomy is associated with the ability of a vehicle to determine its operational environment, thus modulating its behaviour according to relevant norms, needs or constraints (Danks & London, Citation2017).

2 Information regarding the companies with a testing permit in California, the reports used in this study, and the lexicons is available in the dataset stored in the 4TU. Center for Research Data in doi:10.4121/13348535 (Martins Martinho Bessa, Chorus, Kroesen, & Herber, Citation2020).

3 Waymo LLC; Tesla Motors; Nissan; BMW; Ford; Valeo North America Inc.; AutoX Technologies Inc.; Nuro Inc.; Apple Inc.; TuSimple; Aurora Innovation; Toyota Research Institute; Intel Corp; TORC Robotics Inc.; EasyMile; Ridecell; Mercedes Benz; Bosch; GM Cruise LLC; Honda; Zoox Inc.; NVIDIA Corporation; Navya Inc.; Udelv; Pony.AI; Continental Automotive Systems; Mando America Corporation; Uber Advanced Technologies Group; and AImotive Inc.

4 RSS stands for Responsibility-Sensitive Safety and NHTSA stands for National Highway Traffic Safety Administration.

References