4,456
Views
10
CrossRef citations to date
0
Altmetric
Introduction

Unsavory medicine for technological civilization: Introducing ‘Artificial Intelligence & its Discontents’

ORCID Icon

This is, once again, the Age of Artificial Intelligence (Garvey Citation2018c). AI, the suite of techniques intended to make machines capable of performing tasks considered ‘intelligent’ when performed by people, is an epochal technology now colonizing an increasing number of domains, from Internet search and social media to the natural and social sciences; agriculture, banking, criminal sentencing, decision-making, and beyond. AI may soon become ubiquitous, coextensive with technological civilization itself: a taken-for-granted feature of modernity like running water or electricity.

But this does not mean that all is well. While AI promises to liberate and empower users, improve well-being, support social institutions, and enable sustainable development, it also threatens to automate and entrench precarity and illiberalism, degrade mental health, and accelerate the Earth’s ecological collapse.

Freud (Citation1961) famously observed that civilization, despite being ostensibly intended to protect humanity from misery, is paradoxically a great source of unhappiness. Similarly, AI is both touted as the solution to humanity’s biggest problems and decried as one of the biggest problems humankind has ever faced – even, perhaps, its last. A plethora of pundits posit that AI poses an existential risk to human survival on this planet: If not nuclear war, climate catastrophe, or another global pandemic, then it will be ‘superintelligent’ machines that herald the Apocalypse (Barrat Citation2013; Bostrom Citation2014; Clark Citation2014; Yampolskiy Citation2015; Müller Citation2016; Cava Citation2018; Russell Citation2019). Or not. Other AI advocates claim a new wave of ethical developments will usher in the ‘Good AI Society’ (Floridi et al. Citation2018), free from scarcity and strife, thus bringing the West, as Japanese technologist Akihito Kodama (Citation2016) has argued, to its teleological zenith: the return to Eden – abundance without work, life without pain – albeit digitized (Hilton Citation1964; Noble Citation1999; Geraci Citation2010; Diamandis and Kotler Citation2012).

Dystopian hellscape or Edenic utopia? Surely neither of these extremes are the only possible consequence of this vast, sociotechnical system of data, people, places, and things we call ‘AI’ – but what are the alternatives, and where are they to be found when expert partisans on each side dispute both the claims and the credentials of their counterparts? Who can help society make sense of the controversial technoscience of AI? The technoscientists whose careers depend on the success of AI? The business people who employ them? The policymakers devoted to the profits promised by the AI-powered ‘Fourth Industrial Revolution’ (World Economic Forum Citation2016; Schwab Citation2017; Mak Citationn.d.; Brynjolfsson and McAfee Citation2012, Citation2016; McAfee and Brynjolfsson Citation2017; Brynjolfsson and Mitchell Citation2017; Ford Citation2009, Citation2015; though see also Wiener Citation1989; Gimpel Citation1977; Jenkins and Sherman Citation1979; White Citation1980; Johannessen Citation2019)?

None other than the critics. Little sense can be made of AI without reference to its discontents – those who doubt, question, challenge, reject, reform, and otherwise reprise ‘AI’ as it is practiced, promoted, and (re)produced. With the hope of scaffolding deeper, more nuanced understandings of both the epochal transformations being wrought by AI technologies and the range of responses required, possible, and as-yet unimagined, this special issue brings together critical accounts of AI and its discontents, past and present, in order to capture the significance of this historical moment, expand the horizons of the possible, and catalyze sociotechnical action on behalf of diverse publics and future generations whose autonomy – and humanity – are at stake.

AI criticism: the tradition of discontent

Often defined tongue-in-cheek by practitioners as ‘what computers can’t do, yet’ (Hendler and Mulvehill Citation2016), in its relentless focus on future horizons of technical capability, AI is amnesiac – scarcely aware of its own history, much less that of its critics. This may be a consequence of the fact that AI’s history has been written primarily by insiders and developers themselves (McCorduck Citation1979, Citation2004, Citation2019; Crevier Citation1993; Brooks Citation1999; Boden Citation1996; Nilsson Citation2010), all of whom tell triumphant stories of progress towards the current pinnacle upon which the world now stands, with a few bumps along the road thrown in for good measure.

According to this canon, AI began in 1956 at the Dartmouth Conference (Kline Citation2011), and there have only been two ‘bumps’ worth noting ever since – both, coincidentally, professors of philosophy at the University of California, Berkeley: Hubert Dreyfus (Citation1965, Citation1972, Citation1992, Citation2007; Dreyfus and Dreyfus Citation1986, Citation1988), whose phenomenological attack on the Platonic formalism of early AI confounded its pioneers and presciently anticipated their failures; and John Searle (Citation1980, Citation1999, Citation2014; Searle and Kurzweil Citation1999; Denton et al. Citation2002), whose ‘Chinese Room’ thought experiment ‘badly shook the little world of [AI] by claiming and proving (so he said) that there was no such thing’ (Motzkin and Searle Citation1989).

In fact, the tradition of AI criticism – to which the articles of this special issue make an important and timely contribution – is older, richer, and more diverse than suggested by the internalist history of AI (Hoos Citation1960a, Citation1960b, Citation1978; Wiener Citation1960; Greenberger Citation1962; Michael Citation1962; Bureau of Labor Statistics Citation1963; United States Congress Senate Committee on Labor and Public Welfare Citation1963; Neisser Citation1963; Ellul Citation1964; Terborgh Citation1965; Pierce et al. Citation1966; Silberman Citation1966; Hunt Citation1968; Jaki Citation1969; Wheeler Citation1972; Lighthill Citation1973; McDermott Citation1976; Glenn and Feldberg Citation1977; Noble Citation1978; Mori Citation1981; Ornstein, Smith, and Suchman Citation1984, Citation1985; Leontief and Duchin Citation1986; Bloomfield Citation1987; Born Citation1987; Suchman Citation1987; Beusmans and Wieckert Citation1989; Penrose Citation1989; Rosenbrock Citation1989; Brödner Citation1990; Corbett, Rasmussen, and Rauner Citation1991; Ennals Citation1991; Negrotti Citation1991; Collins Citation1992; Maturana and Varela Citation1992; Forsythe Citation1993; Ford Citation2001; Martin Citation1993; Bainbridge et al. Citation1994; Newquist Citation1994; Göranzon Citation1995; Göranzon and Florin Citation1990; Citation1991; Hutchins Citation1995; Edwards Citation1996; Hendriks-Jansen Citation1996; Kling Citation1996; Olazaran Citation1996; Adam Citation1998). Whereas Dreyfus and Searle were lambasted by AI partisans as ignorant outsiders (though see Armstrong, Sotala, and Ó hÉigeartaigh Citation2014), earlier critics, such as the mathematicians Richard Bellman (Citation1958) and Hubert’s brother Stuart (Dreyfus Citation1962, Citation2004, Citation2009, Citation2014), came from within the technical community. Perhaps the most fearsome of these discontents was Mortimer Taube (1911–1965), director of multiple research divisions of the Library of Congress, ‘who, besides being an outstanding theorist and inventor, [was] one of the most successful business practitioners of the computer-based, data-processing art’ (Solo Citation1963, 173; see also Smith Citation1993). Whereas Dreyfus, Searle, and many subsequent critics stayed squarely on theoretical terrain, Taube’s now forgotten masterpiece, Computers and Common Sense: The Myth of Thinking Machines (Citation1961), critiqued the social irresponsibility and economic profligacy of AI as well as its flawed philosophical foundations. Decades before ‘deconstruction’ became a term of art in academe, Taube analysed the AI literature to reveal its hidden assumptions, gaps in reasoning, and antiquated worldviews. Prior to Berger and Luckmann (Citation1966), he showed how AI pioneers used language to socially construct the reality of ‘thinking machines’ through an interlocking web of peer-citation, long before theorists of the actor–network showed this to be a fundamental aspect of technoscientific power (Callon, Law, and Rip Citation1986). As it would become clear shortly after Taube’s sudden death in 1965, the AI pioneers were defrauding everyone, from their military funders to their peers and the broader public, with overhyped claims about their machines. Yet Taube distinguished them from simple criminals by noting that while the creators of sophisticated literary and artistic forgeries are ostensibly aware that they are committing fraud, the creators of ‘thinking machines’ apparently believed they were actually doing science.

Dreyfus famously attacked the scientific credentials of AI by equating it to alchemy (Citation1965). Like Darwin’s Wallace, he arrived at this conclusion independently of Taube, who had argued several years prior that AI was a ‘scientific aberration’ like astrology or physiognomy (Citation1961, 118–128). Taube looked forward to the day when its central dogma, the doctrine of ‘Man-Machine Identity’ – that the human brain is ‘nothing but’ a machine, and therefore can be simulated by another machine (76) – would not only be rejected, but, like the eugenics of biology or the colonial origins of anthropology, disowned entirely as an embarrassment to science. Understanding technoscience as a social activity that is ultimately meaningless if not helpful to society at large, Taube decried AI pioneers for using the term ‘science’ to ‘peddle nostrums to a gullible public’ while avoiding scrutiny by ‘insisting on the pure scientific nature of their intentions’ (124). To counteract their impairing influence on society, Taube attempted to introduce the criticism of technoscience ‘as an enterprise similar in its aims to the established arts of literary, musical, art, and religious criticism,’ one that ‘views the [techno]scientific enterprise as an activity carried out by men [sic], not by demigods, nor even high priests’ (v).

Widely influential at the time of its publication, the significance of Taube’s critical project – which provides an excellent frame and sets a high bar for this special issue – was recognized by no less a technoscientist than Alvin M. Weinberg, Director of Oak Ridge National Laboratory, who observed that while the ‘arts have always taken art critics and art criticism for granted,’ technoscientists typically assume they have no need for critics:

Bad science is science that does not agree with nature; there are, in principle, objective criteria for deciding between good and bad science. But Taube's main contention is that in a field such as [AI] which deals with human artifacts (computers) and with logical, not empirical, issues, the tried-and-true criterion of agreement with experiment no longer serves to cull the bad from the good. Nor is the review of editors or fellow workers or government administrators sufficient—in Taube’s opinion all are tainted with the same poison and, being taken in by the same alleged scientific fraud, can criticize only in detail, not in principle. If the scientific activities Taube criticizes were cheap, not much harm would be done; but since computers (like so much of modern Big Science) are expensive and are supported by public money, Taube argues that it is necessary and valid to subject these activities as a whole to the kind of criticism to which art is subjected, to criticize broadly the essential validity of the enterprise rather than to argue about the details within an accepted conceptual framework. That such a course is excruciatingly difficult, if for no other reason than that science is done by specialists and broad criticism of science must of necessity be done by people who know less than the specialists, does not deter Taube; he sees his duty and he states his opinions without pulling punches. (Weinberg Citation1962, 310)

Weinberg concluded by stating his hope that Taube’s critique would go on to have influence far beyond the narrow confines of AI, for ‘Much of modern Big Science could be helped by a dose of such unsavory, but necessary, medicine.’

Having edited and assembled the 10 articles in this special issue, I can assure the reader that each does its part to bring the bold critical enterprise begun by Taube into the twenty-first century – and that not a single one pulls its punches. Now allow me to echo Weinberg in hoping the contribution of these discontents proves broadly influential, for much of modern technoscience – and civilization itself – could be helped by the unsavory, but necessary, medicine offered herein.

Overview of the issue

In ‘The Lamp and the Lighthouse,’ Zachary Loeb examines the career of one of AI’s most notorious discontents, computer scientist Joseph Weizenbaum (Citation1976), the first of a handful of notable defectors (e.g. Winograd and Flores Citation1987; Agre Citation1997; Suchman Citation2007; Marcus and Davis Citation2019; Smith Citation2019) from what he called the ‘Artificial Intelligentsia.’ Delving into Weizenbaum’s correspondence with the historian Lewis Mumford, whose own work affords considerable cultural context on the development of AI (Mumford Citation1963, Citation1964, Citation1965, Citation1967, Citation1970), Loeb provides an important corrective to the official canon: Weizenbaum was not – as so often portrayed by his former colleagues and adversaries in AI – a lone wolf, howling in the wilderness. Rather, his work, and the tradition of AI discontent more generally, was part of a larger tradition of social criticism responding to the accelerating automation, computerization, and complexification of technological civilization in the twentieth century.

Depending upon how ‘AI’ is defined, however, discontent is ancient (Wiener Citation1964; Cohen Citation1966; Winner Citation1989; Noble Citation1999; Herzfeld Citation2002; Geraci Citation2010; Russell and Norvig Citation2010; Garfinkel and Grunspan Citation2018). By excavating a minor literature on ‘Artificial Stupidity,’ Michael Falk’s article extends the critical tradition of the discontented beyond well-trod tomes like Forster’s ‘The Machine Stops’ (Citation1909), the apocryphal ‘Book of the Machines’ in Butler’s Erewhon (Citation1872), as well as Shelley’s 1818 Frankenstein, into the seventeenth and eighteenth centuries. Inverting narratives about the potential dangers of a hypothetical ‘superintelligence,’ Falk explicates the real risks of actual machine stupidity to open up a new (old) line of inquiry: Instead of always probing whether a given machine is truly intelligent – whatever that means – we ought instead to enquire, ‘What kind of stupid is it?’

Somewhat surprisingly, there is little if any consideration of stupidity in AI. But even more surprisingly, there is scarcely any more attention paid to what would seem to be the central concern of the entire AI enterprise: intelligence. Harry Collins, a sociologist and longtime discontent who critiques AI from the standpoint of Science and Technology Studies, summarizes and distills much of his oeuvre (Citation1989, Citation1990, Citation1992, Citation2018; Collins and Kusch Citation1998) into ‘The Science of Artificial Intelligence and its Critics.’ It provides an important tool for puncturing contemporary versions of the ‘myth of thinking machines’: a six-level scale of intelligence that clarifies what it actually means to be intelligent from a sociological perspective. In addition to combating hype by helping ordinary people distinguish fact from fiction in AI, Collins proposes that his framework could be used by experts to make AI a respectable science again – but do they have the ears to hear his ‘productive criticism’?

As Taube pointed out decades ago, hype-busting is important because the contemporary AI enterprise is not, and has never been, just a bit of harmless technoscientific experimentation (Garvey Citation2018a; Garvey and Maskal Citation2019). With its renewed geostrategic importance (Lee Citation2018; Scharre Citation2018; Comiter Citation2019; Garvey Citation2019b; Lin Citation2019; Mecklin Citation2019; NSCAI Citation2019; Prakash Citation2019; Bulletin of the Atomic Scientists Citation2020; Johnson Citation2020) and increasing presence in nearly every sector of society (Citron and Pasquale Citation2014; Pasquale Citation2015; Pasquinelli Citation2015; O’Neil Citation2016; Eubanks Citation2017; Tegmark Citation2017; Shoham et al. Citation2017, Citation2018; Wachter-Boettcher Citation2017; Broussard Citation2018; Foer Citation2018; S. U. Noble Citation2018; Taplin Citation2018; Susskind Citation2018; Vaidhyanathan Citation2018; Zuboff Citation2018; Atanasoski and Vora Citation2019; Benjamin Citation2019; Frey Citation2019; Perrault et al. Citation2019; Topol Citation2019; Webb Citation2019; Nourbakhsh and Keating Citation2020), enormous national budgets have been and are being planned on the promise that more AI will be ‘good’ for society (Webster et al. Citation2017; Dutton, Barron, and Boskovic Citation2018; The White House Citation2018; European Commission and Joint Research Centre Citation2018; Schmidt et al. Citation2020). Yet as Ulnicane and colleagues demonstrate in ‘Good Governance as a Response to Discontents? Déjà vu, or Lessons for AI from other Emerging Technologies,’ the ‘AI for Good’ narrative at the centre of these (inter)national initiatives is ahistorical, reproducing the amnesia of the AI canon by ignoring and resisting multiple relevant precedents in the governance of emerging technologies, such as public engagement and responsible innovation (Jasanoff Citation1996; Rowe and Frewer Citation2005; Wynne Citation2006; Stilgoe, Lock, and Wilsdon Citation2014; Özdemir and Springer Citation2018; Garvey Citation2019a).

Cheryl Holzmeyer’s ‘Beyond ‘AI for Social Good’ (AI4SG): Social Transformations – Not Tech-Fixes – for Health Equity,’ the fifth article in our special issue, similarly shows that these initiatives, however well-intentioned, breed discontent by distracting from root causes of social inequity and the meliorative potential of challenging existing systems of power. While advocates claim AI will improve public health, perhaps by making it possible to predict ‘well-being’ at the population level (Jaidka et al. Citation2020) or rapidly screen women for breast cancer (McKinney et al. Citation2020), all sociotechnical failures aside (Herper Citation2017; Ross Citation2018; Strickland Citation2019), the technical community’s narrow focus on potential downstream interventions contributes to neglect of the social determinants of health upstream, such as adequate income and housing, which in turn amplifies inequality and puts more of society at risk. In other words, although AI is supposed to be a powerful tool for improving well-being, by diverting attention and resources away from fundamentals into expensive, experimental, expert-driven technical systems, it is paradoxically one of the greatest potential sources of social harm.

The next two articles expand upon the medical theme by exploring the role of AI in the clinic. In ‘Don’t Touch My Stuff: Historicizing Resistance to AI and Algorithmic Computer Technologies in Medicine,’ Ariane Hanemaayer documents the long tradition of doctors’ discontent as a struggle with and against machines over the authorship of medical truth. Rather than continuing to contest AI unaided, Hanemaayer suggests that physicians might find allies in their patients, with whom they share a partisan interest in reducing and protecting against the biases and other potential harms of medical AI systems (Coiera Citation1996; Cabitza, Rasoini, and Gensini Citation2017; Char, Shah, and Magnus Citation2018; Garvey Citation2018b; Krittanawong Citation2018). Saheli Datta Burton, Tara Mahfoud, and colleagues’ ‘Clinical Translation of Computational Brain Models: Understanding the Salience of Trust in Clinician-Researcher Relationships’ explores the physician’s predicament from another angle – the quintessential social bond of trust (Yamagishi Citation2011; Adali Citation2013). Drawing on their extensive experience with the Human Brain Project and interviews with experts, they show that without gaining clinicians’ trust through upstream collaboration that builds upon practitioners’ tacit knowledge, medical AI systems are unlikely to make meaningful contributions to patients’ health, even if they are adopted into the clinic.

The last three articles expand the frame of discontent to include the issues of biology, humanity, and identity. In ‘Truth from the Machine: Artificial Intelligence and the Materialisation of Identity,’ Keyes, Hitzig, and Blell explore how the natural sciences change when investigators utilize the quantitative techniques of AI to ‘discover’ qualitative social constructs such as ‘disease’ and ‘sexuality.’ Rachel Adams, in asking ‘Can Artificial Intelligence Be Decolonised?,’ authoritatively unpacks the intertwined legacies of colonialism, racism, and Western cultural hegemony that underlie the conceptual foundations of AI, in order to establish an erudite theorization of what the strategy of decoloniality must mean for the field if it is not to be (re)appropriated as yet another ‘ethic.’ And finally, Alan Blackwell offers a personal account of his own discontent as a longtime practitioner that builds into a proposal for a future ethnography that would make it possible to think and conduct AI otherwise.

The future of AI and its discontents

Some discontents come from outside the field, others from within. Most if not all of them, however, critique AI in order to address larger issues of continued cultural relevance, such as the nature of ‘intelligence’; the development, implementation, and governance of large-scale sociotechnical systems (Garvey Citation2018d); the consequences of doing AI within the military–industrial–university complex; the problems of mind, brain, and consciousness in a material universe; the relationship between language, thought, and society; as well as what it means to be ‘human’ in an increasingly computerized world.

These criticisms from the discontented reveal the deceptively simple two-letter moniker ‘AI’ to be a microcosm of technological civilization in dire need of strong medicine. While no single dose is strong enough to serve as an antidote to the cyclical malaise of modern machine-driven madness (Garvey Citation2018c), as AI grows more pervasive, its discontents will grow more numerous, and their critical prescriptions, however unsavory, ever more important to heed.

Acknowledgments

Shunryu Colin Garvey would like to thank the editor of ISR, Willard McCarty, for his patience, wisdom, and support on this special issue; Alan Blackwell for reading an early draft; Ariane Hanemaayer for her close collaboration and continued dedication to the project of ‘AI & its Discontents’; and all the contributors to this issue, as well as AI discontents past, present, and future.

Disclosure statement

No potential conflict of interest was reported by the author(s).

Additional information

Notes on contributors

Shunryu Colin Garvey

Dr. Shunryu Colin Garvey is trained in the interdisciplinary field of Science & Technology Studies (STS). He uses AI as a case study to probe decision making under the conditions of complexity, uncertainty, and disagreement, in order understand how societies can more safely, fairly, and wisely govern controversial sociotechnical systems.

References

  • Adali, Sibel. 2013. Modeling Trust Context in Networks. SpringerBriefs in Computer Science. New York, NY: Springer.
  • Adam, Alison. 1998. Artificial Knowing: Gender and the Thinking Machine. New York, NY: Routledge.
  • Agre, Philip. 1997. Computation and Human Experience. Learning in Doing: Social, Cognitive and Computational Perspectives. Cambridge: Cambridge University Press.
  • Armstrong, Stuart, Kaj Sotala, and Seán S. Ó hÉigeartaigh. 2014. “The Errors, Insights and Lessons of Famous AI Predictions – and What They Mean for the Future.” Journal of Experimental & Theoretical Artificial Intelligence 26 (3): 317–342. https://doi.org/https://doi.org/10.1080/0952813X.2014.895105.
  • Atanasoski, Neda, and Kalindi Vora. 2019. Surrogate Humanity: Race, Robots, and the Politics of Technological Futures. Perverse Modernities. Durham, NC: Duke University Press.
  • Bainbridge, William Sims, Edward E. Brent, Kathleen M. Carley, David R. Heise, Michael W. Macy, Barry Markovsky, and John Skvoretz. 1994. “Artificial Social Intelligence.” Annual Review of Sociology 20 (1): 407–436. https://doi.org/https://doi.org/10.1146/annurev.so.20.080194.002203.
  • Barrat, James. 2013. Our Final Invention: Artificial Intelligence and the End of the Human Era. New York, NY: Thomas Dunne Books.
  • Bellman, Richard. 1958. “On ‘Heuristic Problem Solving’ by Simon and Newell.” Operations Research 6 (3): 448–449.
  • Benjamin, Ruha. 2019. Race After Technology: Abolitionist Tools for the New Jim Code. Medford, MA: Polity.
  • Berger, Peter L., and Thomas Luckmann. 1966. The Social Construction of Reality: A Treatise in the Sociology of Knowledge. 1st ed. Garden City, NY: Doubleday.
  • Beusmans, Jack, and Karen Wieckert. 1989. “Computing, Research, and War: If Knowledge Is Power, Where Is Responsibility?” Communications of the ACM 32 (8): 939–951. https://doi.org/https://doi.org/10.1145/65971.65973.
  • Bloomfield, Brian P. 1987. The Question of Artificial Intelligence: Philosophical and Sociological Perspectives. New York, NY: Croom Helm.
  • Boden, Margaret A., ed. 1996. Artificial Intelligence. 1st ed. San Diego, CA: Academic Press.
  • Born, Rainer P., ed. 1987. Artificial Intelligence: The Case Against. New York, NY: St. Martin’s Press.
  • Bostrom, Nick. 2014. Superintelligence: Paths, Dangers, Strategies. Oxford: Oxford University Press.
  • Brödner, Peter. 1990. The Shape of Future Technology: The Anthropocentric Alternative. The Springer Series on Artificial Intelligence and Society. London: Springer London. https://doi.org/https://doi.org/10.1007/978-1-4471-1733-9.
  • Brooks, Rodney Allen. 1999. Cambrian Intelligence: The Early History of the New AI. Cambridge, MA: MIT Press.
  • Broussard, Meredith. 2018. Artificial Unintelligence: How Computers Misunderstand the World. Cambridge, MA: The MIT Press.
  • Brynjolfsson, Erik, and Andrew McAfee. 2012. Race Against the Machine: How the Digital Revolution Is Accelerating Innovation, Driving Productivity, and Irreversibly Transforming Employment and the Economy. Lexington, MA: Digital Frontier Press.
  • Brynjolfsson, Erik, and Andrew McAfee. 2016. The Second Machine Age: Work, Progress, and Prosperity in the Time of Brilliant Technologies. New York; London: W. W. Norton.
  • Brynjolfsson, Erik, and Tom Mitchell. 2017. “What Can Machine Learning Do? Workforce Implications.” Science 358 (6370): 1530–1534.
  • Bulletin of the Atomic Scientists. 2020. “2020 Doomsday Clock Statement.” Edited by John Mecklin. Bulletin of the Atomic Scientists, January, 20.
  • Bureau of Labor Statistics. 1963. “Implications of Automation and Other Technological Developments: A Selected Annotated Bibliography.” 1319–1. US Department of Labor.
  • Butler, Samuel. 1872. Erewhon. London: Trübner.
  • Cabitza, Federico, Raffaele Rasoini, and Gian Franco Gensini. 2017. “Unintended Consequences of Machine Learning in Medicine.” JAMA 318 (6): 517. https://doi.org/https://doi.org/10.1001/jama.2017.7797.
  • Callon, Michel, John Law, and Arie Rip, eds. 1986. Mapping the Dynamics of Science and Technology: Sociology of Science in the Real World. London: The MacMillan Press Ltd.
  • Cava, Marco Della. 2018. “Elon Musk Says AI Could Doom Human Civilization. Zuckerberg Disagrees. Who’s Right?” USA TODAY, January 2, 2018. https://www.usatoday.com/story/tech/news/2018/01/02/artificial-intelligence-end-world-overblown-fears/985813001/.
  • Char, Danton S., Nigam H. Shah, and David Magnus. 2018. “Implementing Machine Learning in Health Care – Addressing Ethical Challenges.” New England Journal of Medicine 378 (11): 981–983. https://doi.org/http://doi.org/10.1056/NEJMp1714229.
  • Citron, Danielle Keats, and Frank A. Pasquale. 2014. “The Scored Society: Due Process for Automated Predictions.” Washington Law Review 89 (1): 1–33.
  • Clark, Stuart. 2014. “Artificial Intelligence Could Spell End of Human Race – Stephen Hawking.” The Guardian, December 2, sec. Science. https://www.theguardian.com/science/2014/dec/02/stephen-hawking-intel-communication-system-astrophysicist-software-predictive-text-type.
  • Cohen, John. 1966. Human Robots in Myth and Science. London: George Allen & Unwin Ltd.
  • Coiera, Enrico W. 1996. “Artificial Intelligence in Medicine: The Challenges Ahead.” Journal of the American Medical Informatics Association 3 (6): 363–366. https://doi.org/https://doi.org/10.1136/jamia.1996.97084510.
  • Collins, Harry M. 1989. “Computers and the Sociology of Scientific Knowledge.” Social Studies of Science 19 (4): 613–624.
  • Collins, Harry M. 1990. Artificial Experts: Social Knowledge and Intelligence Machines. Cambridge, MA: The MIT Press.
  • Collins, Harry M. 1992. “Hubert L. Dreyfus, Forms of Life, and a Simple Test for Machine Intelligence.” Social Studies of Science 22 (4): 726–739.
  • Collins, Harry M. 2018. Artifictional Intelligence: Against Humanity’s Surrender to Computers. Cambridge: Polity Press.
  • Collins, Harry M., and Martin Kusch. 1998. The Shape of Actions: What Humans and Machines Can Do. Cambridge, MA: MIT Press.
  • Collins, Randall. 1992. “Can Sociology Create an Artificial Intelligence?” In Sociological Insight: An Introduction to Non-Obvious Sociology, 155–188. New York, NY: Oxford University Press.
  • Comiter, Marcus. 2019. Attacking Artificial Intelligence: AI’s Security Vulnerability and What Policymakers Can Do About It. Cambridge, MA: Harvard Kennedy School, Belfer Center for Science and International Affairs.
  • Corbett, J. Martin, Lauge Baungaard Rasmussen, and Felix Rauner. 1991. Crossing the Border: The Social and Engineering Design of Computer Integrated Manufacturing Systems. The Springer Series on Artificial Intelligence and Society. London; New York: Springer-Verlag.
  • Crevier, Daniel. 1993. AI: The Tumultuous History of the Search for Artificial Intelligence. New York, NY: Basic Books.
  • Denton, Michael, Thomas Ray, William A Dembski, John R Searle, George Gilder, and Ray Kurzweil. 2002. Are We Spiritual Machines?: Ray Kurzweil vs the Critics of Strong A.I. Edited by Jay W. Richards. 1st ed. Seattle, WA: Discovery Institute Press.
  • Diamandis, Peter H., and Steven Kotler. 2012. Abundance: The Future Is Better Than You Think. 1st Free Press hardcover ed. New York, NY: Free Press.
  • Dreyfus, Stuart. 1962. “Artificial Intelligence: Deus Ex Machina.” The Second Coming, June.
  • Dreyfus, Hubert L. 1965. Alchemy and Artificial Intelligence. P-3244. Santa Monica, CA: RAND Corporation.
  • Dreyfus, Hubert L. 1972. What Computers Can’t Do: A Critique of Artificial Reason. New York, NY: HarperCollins Publishers.
  • Dreyfus, Hubert L. 1992. What Computers Still Can’t Do. New York, NY: MIT Press.
  • Dreyfus, Hubert L. 2007. “Why Heideggerian AI Failed and How Fixing It Would Require Making It More Heideggerian.” Philosophical Psychology 20 (2): 247–268. https://doi.org/https://doi.org/10.1080/09515080701239510.
  • Dreyfus, Hubert L., and Stuart E. Dreyfus. 1986. “Competent Systems: The Only Future for Inference-Making Computers.” Future Generation Computer Systems 2 (4): 233–243. https://doi.org/https://doi.org/10.1016/0167-739X(86)90023-3.
  • Dreyfus, Hubert L., and Stuart E. Dreyfus. 1988. Mind Over Machine: The Power of Human Intuition and Expertise in the Era of the Computer. New York, NY: Free Press.
  • Dreyfus, Stuart E. 2004. “Totally Model-Free Learned Skillful Coping.” Bulletin of Science, Technology & Society 24 (3): 182–187. https://doi.org/https://doi.org/10.1177/0270467604264813.
  • Dreyfus, Stuart E. 2009. “A Modern Perspective on Creative Cognition.” Bulletin of Science, Technology & Society 29 (1): 3–8. https://doi.org/https://doi.org/10.1177/0270467608328708.
  • Dreyfus, Stuart E. 2014. “System 0: The Overlooked Explanation of Expert Intuition.” In Handbook of Research Methods on Intuition, edited by Marta Sinclair, 15–27. Handbooks of Research Methods in Management. Cheltenham; Northampton, MA: Edward Elgar.
  • Dutton, Tim, Brent Barron, and Gaga Boskovic. 2018. Building an AI World: Report on National and Regional AI Strategies. Toronto, ON: CIFAR. https://www.cifar.ca/docs/default-source/ai-society/buildinganaiworld_eng.pdf.
  • Edwards, Paul N. 1996. The Closed World: Computers and the Politics of Discourse in Cold War America. Inside Technology. Cambridge, MA: MIT Press.
  • Ellul, Jacques. 1964. The Technological Society. New York, NY: Vintage Books.
  • Ennals, Richard. 1991. Artificial Intelligence and Human Institutions. The Springer Series on Artificial Intelligence and Society. London; New York, NY: Springer-Verlag.
  • Eubanks, Virginia. 2017. Automating Inequality: How High-Tech Tools Profile, Police, and Punish the Poor. 1st ed. New York, NY: St. Martin’s Press.
  • European Commission, and Joint Research Centre. 2018. “Artificial Intelligence: A European Perspective.” http://publications.europa.eu/publication/manifestation_identifier/PUB_KJNA29425ENE.
  • Floridi, Luciano, Josh Cowls, Monica Beltrametti, Raja Chatila, Patrice Chazerand, Virginia Dignum, Christoph Luetge, et al. 2018. “AI4People—an Ethical Framework for a Good AI Society: Opportunities, Risks, Principles, and Recommendations.” Minds and Machines 28 (4): 689–707. https://doi.org/https://doi.org/10.1007/s11023-018-9482-5.
  • Foer, Franklin. 2018. World Without Mind: The Existential Threat of Big Tech. New York, NY: Penguin Books.
  • Forsythe, Diana E. 2001. Studying Those Who Study Us: An Anthropologist in the World of Artificial Intelligence. Stanford, CA: Stanford University Press.
  • Ford, Martin. 2009. The Lights in the Tunnel: Automation, Accelerating Technology and the Economy of the Future. Acculant Publishing.
  • Ford, Martin. 2015. Rise of the Robots: Technology and the Threat of a Jobless Future. New York, NY: Basic Books.
  • Forster, E. M. 1909. “The Machine Stops.” The Oxford and Cambridge Review, November.
  • Forsythe, Diana E. 1993. “The Construction of Work in Artificial Intelligence.” Science, Technology & Human Values 18 (4): 460–479.
  • Freud, Sigmund. 1961. Civilization and Its Discontents. New York, NY: W. W. Norton.
  • Frey, Carl Benedikt. 2019. The Technology Trap: Capital, Labor, and Power in the Age of Automation. Princeton, NJ: Princeton University Press.
  • Garfinkel, Simson, and Rachel H Grunspan. 2018. The Computer Book: From the Abacus to Artificial Intelligence, 250 Milestones in the History of Computer Science. https://www.overdrive.com/search?q=62FCFE29-6B04-40AC-BFA4-3FFCD565784B.
  • Garvey, Colin. 2018a. “AI Risk Mitigation Through Democratic Governance: Introducing the 7-Dimensional AI Risk Horizon.” In Proceedings of the 2018 AAAI/ACM Conference on AI, Ethics, and Society, 366–367. AIES ‘18. New York, NY: ACM. https://doi.org/https://doi.org/10.1145/3278721.3278801.
  • Garvey, Colin. 2018b. “Interview with Colin Garvey, Rensselaer Polytechnic Institute. Artificial Intelligence and Systems Medicine Convergence.” OMICS: A Journal of Integrative Biology 22 (2): 130–132. https://doi.org/https://doi.org/10.1089/omi.2017.0218.
  • Garvey, Colin. 2018c. “Broken Promises and Empty Threats: The Evolution of AI in the USA, 1956-1996.” Technology’s Stories, March. https://doi.org/https://doi.org/10.15763/jou.ts.2018.03.16.18.
  • Garvey, Colin. 2018d. “ A Framework for Evaluating Barriers to the Democratization of Artificial Intelligence.” In Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence, New Orleans, Louisiana USA - February 2-7, 2018, 8079-8080. Palo Alto, CA: AAAI Press. https://aaai.org/ocs/index.php/AAAI/AAAI18/paper/view/17320/16477.
  • Garvey, Colin. 2019a. “Hypothesis: Is ‘Terminator Syndrome’ a Barrier to Democratizing Artificial Intelligence and Public Engagement in Digital Health?” OMICS: A Journal of Integrative Biology 23 (7): 362–363. https://doi.org/https://doi.org/10.1089/omi.2019.0070.
  • Garvey, Colin. 2019b. “Artificial Intelligence and Japan’s Fifth Generation: The Information Society, Neoliberalism, and Alternative Modernities.” Pacific Historical Review 88 (4): 619–658. https://doi.org/https://doi.org/10.1525/phr.2019.88.4.619.
  • Garvey, Colin, and Chandler Maskal. 2019. “Sentiment Analysis of the News Media on Artificial Intelligence Does Not Support Claims of Negative Bias Against Artificial Intelligence.” OMICS: A Journal of Integrative Biology 23: 1–16. https://doi.org/https://doi.org/10.1089/omi.2019.0078.
  • Geraci, Robert M. 2010. Apocalyptic AI: Visions of Heaven in Robotics, Artificial Intelligence, and Virtual Reality. New York: Oxford University Press.
  • Gimpel, Jean. 1977. The Medieval Machine: The Industrial Revolution of the Middle Ages. New York: Penguin Books.
  • Glenn, Evelyn Nakano, and Roslyn L. Feldberg. 1977. “Degraded and Deskilled: The Proletarianization of Clerical Work.” Social Problems 25 (1): 52–64. https://doi.org/https://doi.org/10.2307/800467.
  • Göranzon, Bo, ed. 1995. Skill, Technology, and Enlightenment: On Practical Philosophy. The Springer Series on Artificial Intelligence and Society. London; New York: Springer-Verlag.
  • Göranzon, Bo, and Magnus Florin, eds. 1990. Artifical Intelligence, Culture and Language: On Education and Work. The Springer Series on Artificial Intelligence and Society. London: Springer London. https://doi.org/https://doi.org/10.1007/978-1-4471-1729-2.
  • Göranzon, Bo, and Magnus Florin, eds. 1991. Dialogue and Technology: Art and Knowledge. The Springer Series on Artificial Intelligence and Society. London; New York: Springer-Verlag.
  • Greenberger, Martin. 1962. Computers and the World of the Future. Cambridge, MA: MIT Press.
  • Hendler, James, and Alice M. Mulvehill. 2016. Social Machines: The Coming Collision of Artificial Intelligence, Social Networking, and Humanity. Berkeley, CA: Apress.
  • Hendriks-Jansen, Horst. 1996. Catching Ourselves in the Act: Situated Activity, Interactive Emergence, Evolution, and Human Thought. Complex Adaptive Systems. Cambridge, MA: MIT Press.
  • Herper, Matthew. 2017. “MD Anderson Benches IBM Watson in Setback For Artificial Intelligence in Medicine.” Forbes, February 19. https://www.forbes.com/sites/matthewherper/2017/02/19/md-anderson-benches-ibm-watson-in-setback-for-artificial-intelligence-in-medicine/.
  • Herzfeld, Noreen L. 2002. In Our Image: Artificial Intelligence and the Human Spirit. Theology and the Sciences. Minneapolis, MN: Fortress Press.
  • Hilton, Alice Mary. 1964. “Cyberculture – The Age of Abundance and Leisure.” Michigan Quarterly Review 3 (4). http://hdl.handle.net/2027/spo.act2080.0003.004:03.
  • Hoos, Ida R. 1960a. “The Impact of Office Automation on Workers.” International Labour Review 82 (4): 363–388.
  • Hoos, Ida R. 1960b. “The Sociological Impact of Automation in the Office.” Management Science MT-1 (2): 10–19. https://doi.org/https://doi.org/10.1287/mantech.1.2.10.
  • Hoos, Ida R. 1978. “Paperwork Control.” Society 16 (1): 5–8. https://doi.org/https://doi.org/10.1007/BF02712550.
  • Hunt, E. 1968. “Computer Simulation: Artificial Intelligence Studies and Their Relevance to Psychology.” Annual Review of Psychology 19 (1): 135–168. https://doi.org/https://doi.org/10.1146/annurev.ps.19.020168.001031.
  • Hutchins, Edwin. 1995. Cognition in the Wild. Cambridge, MA: MIT Press.
  • Jaidka, Kokil, Salvatore Giorgi, H. Andrew Schwartz, Margaret L. Kern, Lyle H. Ungar, and Johannes C. Eichstaedt. 2020. “Estimating Geographic Subjective Well-Being from Twitter: A Comparison of Dictionary and Data-Driven Language Methods.” Proceedings of the National Academy of Sciences 117 (19): 10165–10171. https://doi.org/https://doi.org/10.1073/pnas.1906364117.
  • Jaki, Stanley L. 1969. Brains, Minds, and Computers. New York, NY: Herder and Herder, Inc.
  • Jasanoff, Sheila. 1996. “Beyond Epistemology: Relativism and Engagement in the Politics of Science.” Social Studies of Science 26 (2): 393–418.
  • Jenkins, Clive, and Barrie Sherman. 1979. The Collapse of Work. London: Eyre Methuen.
  • Johannessen, Jon-Arild. 2019. The Workplace of the Future: The Fourth Industrial Revolution, the Precariat and the Death of Hierarchies. Routledge Studies in the Economics of Innovation. Abingdon, Oxon; New York, NY: Routledge, Taylor & Francis Group.
  • Johnson, James S. 2020. “Artificial Intelligence: A Threat to Strategic Stability.” Strategic Studies Quarterly 24: 16–39.
  • Kline, Ronald. 2011. “Cybernetics, Automata Studies, and the Dartmouth Conference on Artificial Intelligence.” IEEE Annals of the History of Computing 33 (4): 5–16. https://doi.org/https://doi.org/10.1109/MAHC.2010.44.
  • Kling, Rob. 1996. Computerization and Controversy: Value Conflicts and Social Choices. San Diego, CA: Morgan Kaufmann.
  • Kodama, Akihito. 2016. Jinkō chinō wa watashitachi wo horobosunoka: keisanki ga kami ni naru 100 nen no monogatari [Will AI Destroy us all? The 100 Year Story of Calculators Becoming Gods]. Tokyo: Diamond Publishers.
  • Krittanawong, C. 2018. “The Rise of Artificial Intelligence and the Uncertain Future for Physicians.” European Journal of Internal Medicine 48 (February): e13–e14. https://doi.org/https://doi.org/10.1016/j.ejim.2017.06.017.
  • Lee, Kai-Fu. 2018. AI Superpowers: China, Silicon Valley, and the New World Order. Boston, MA: Houghton Mifflin Harcourt.
  • Leontief, Wassily, and Faye Duchin. 1986. The Future Impact of Automation on Workers. New York, NY: Oxford University Press.
  • Lighthill, James. 1973. Lighthill Report: Artificial Intelligence. London: Science Research Council (SRC).
  • Lin, Herbert. 2019. “The Existential Threat from Cyber-Enabled Information Warfare.” Bulletin of the Atomic Scientists 75 (4): 187–196. https://doi.org/https://doi.org/10.1080/00963402.2019.1629574.
  • Mak, Alan. n.d. Getting to the Future First: How Britain Can Lead the Fourth Industrial Revolution. London: ConservativeHome: the home of conservativism.
  • Marcus, Gary, and Ernest Davis. 2019. Rebooting AI: Building Artificial Intelligence We Can Trust. 1st ed. New York, NY: Pantheon Books.
  • Martin, C. Dianne. 1993. “The Myth of the Awesome Thinking Machine.” Communications of the ACM 36 (4): 120–133. https://doi.org/https://doi.org/10.1145/255950.153587.
  • Maturana, Humberto R., and Francisco J. Varela. 1992. The Tree of Knowledge: The Biological Roots of Human Understanding. Rev. ed. Boston, MA; New York, NY: Shambhala; Distributed in the U.S. by Random House.
  • McAfee, Andrew, and Erik Brynjolfsson. 2017. Machine, Platform, Crowd: Harnessing Our Digital Future. New York, NY: W. W. Norton & Company.
  • McCorduck, Pamela. 1979. Machines Who Think: A Personal Inquiry Into the History and Prospects of Artificial Intelligence. San Francisco, CA: W. H. Freeman.
  • McCorduck, Pamela. 2004. Machines Who Think: A Personal Inquiry Into the History and Prospects of Artificial Intelligence. 25th anniversary update. Natick, MA: A.K. Peters.
  • McCorduck, Pamela. 2019. This Could Be Important: My Life and Times with the Artificial Intelligentsia. Pittsburgh, PA: Carnegie Mellon University: ETC Press: Signature.
  • McDermott, Drew. 1976. “Artificial Intelligence Meets Natural Stupidity.” ACM SIGART Bulletin 57 (April): 4–9. https://doi.org/https://doi.org/10.1145/1045339.1045340.
  • McKinney, Scott Mayer, Marcin Sieniek, Varun Godbole, Jonathan Godwin, Natasha Antropova, Hutan Ashrafian, Trevor Back, et al. 2020. “International Evaluation of an AI System for Breast Cancer Screening.” Nature 577 (7788): 89–94. https://doi.org/https://doi.org/10.1038/s41586-019-1799-6.
  • Mecklin, John. 2019. “Dealing Realistically with the Artificial Intelligence Revolution.” Bulletin of the Atomic Scientists 75 (3): 93–94. https://doi.org/https://doi.org/10.1080/00963402.2019.1604812.
  • Michael, Donald N. 1962. Cybernation: The Silent Conquest. Santa Barbara, CA: Center for the Study of Democratic Institutions.
  • Mori, Masahiro. 1981. The Buddha in the Robot. 1st. English ed. Tokyo: Kosei Pub. Co.
  • Motzkin, Elhanan, and John R. Searle. 1989. “Artificial Intelligence and the Chinese Room: An Exchange.” The New York Review of Books, February 16. http://www.nybooks.com/articles/1989/02/16/artificial-intelligence-and-the-chinese-room-an-ex/.
  • Müller, Vincent C.2016. Risks of Artificial Intelligence. Boca Raton, FL: CRC Press.
  • Mumford, Lewis. 1963. Technics and Civilization. 2nd ed. New York, NY: Harcourt, Brace & World, Inc.
  • Mumford, Lewis. 1964. “Authoritarian and Democratic Technics.” Technology and Culture 5 (1): 1–8.
  • Mumford, Lewis. 1965. “Utopia, The City and The Machine.” In Utopias and Utopian Thought, edited by Frank E. Manuel, 3–24. Boston, MA: Beacon Press.
  • Mumford, Lewis. 1967. Technics and Human Development. Vol. 1. 2 vols. The Myth of the Machine. New York, NY: Harcourt, Brace & World.
  • Mumford, Lewis. 1970. The Pentagon of Power. Vol. 2. 2 vols. The Myth of the Machine. New York, NY: Harcourt, Brace.
  • Negrotti, Massimo, ed. 1991. Understanding the Artificial: On the Future Shape of Artificial Intelligence. The Springer Series on Artificial Intelligence and Society. London and New York: Springer-Verlag.
  • Neisser, Ulric. 1963. “The Imitation of Man by Machine.” Science 139 (3551): 193–197.
  • Newquist, H. P. 1994. The Brain Makers. 1st ed. Indianapolis, IN: Sams Publishing.
  • Nilsson, Nils J. 2010. The Quest for Artificial Intelligence: A History of Ideas and Achievements. Cambridge; New York: Cambridge University Press.
  • Noble, David F. 1978. “Social Choice in Machine Design: The Case of Automatically Controlled Machine Tools, and a Challenge for Labor.” Politics & Society 8 (3–4): 313–347.
  • Noble, David F. 1999. The Religion of Technology: The Divinity of Man and the Spirit of Invention. New York, NY: Penguin.
  • Noble, Safiya Umoja. 2018. Algorithms of Oppression: How Search Engines Reinforce Racism. New York, NY: New York University Press.
  • Nourbakhsh, Illah Reza, and Jennifer Keating. 2020. AI & Humanity. Cambridge, MA: The MIT Press.
  • NSCAI. 2019. Interim Report. Washington, DC: National Security Commission on Artificial Intelligence.
  • Olazaran, Mikel. 1996. “A Sociological Study of the Official History of the Perceptrons Controversy.” Social Studies of Science 26 (3): 611–659.
  • O’Neil, Cathy. 2016. Weapons of Math Destruction: How Big Data Increases Inequality and Threatens Democracy. London: Allen Lane, Penguin Books.
  • Ornstein, Severo M., Brian C. Smith, and Lucy A. Suchman. 1984. “Strategic Computing.” Bulletin of the Atomic Scientists 40 (10): 11–15. https://doi.org/https://doi.org/10.1080/00963402.1984.11459292.
  • Ornstein, Severo M., Brian C. Smith, and Lucy A. Suchman. 1985. “Strategic Computing: An Assessment.” Communications of the ACM 28 (2): 134–136. https://doi.org/https://doi.org/10.1145/2786.314986.
  • Özdemir, Vural, and Simon Springer. 2018. “What Does ‘Diversity’ Mean for Public Engagement in Science? A New Metric for Innovation Ecosystem Diversity.” OMICS: A Journal of Integrative Biology 22 (3): 184–189. https://doi.org/https://doi.org/10.1089/omi.2018.0002.
  • Pasquale, Frank. 2015. The Black Box Society: The Secret Algorithms That Control Money and Information. Cambridge, MA: Harvard University Press.
  • Pasquinelli, Matteo, ed. 2015. Alleys of Your Mind: Augmented Intelligence and Its Traumas. Lüneburg: Meson press.
  • Penrose, Roger. 1989. The Emperor’s New Mind: Concerning Computers, Minds, and the Laws of Physics. Oxford; New York: Oxford University Press.
  • Perrault, Raymond, Yoav Shoham, Erik Brynjolfsson, Jack Clark, John Etchemendy, Barbara Grosz, Terah Lyons, James Manyika, and Juan Carlos Niebles. 2019. AI Index 2019. Stanford, CA: Stanford Institute for Human-Centered AI.
  • Pierce, John R., John B. Carroll, Eric P. Hamp, David G. Hays, Charles F. Hockett, Anthony G. Oettinger, and Alan Perlis. 1966. Language and Machines: Computers in Translation and Linguistics. 1416. Washington, DC: Automatic Language Processing Advisory Committee, National Academy of Sciences, National Research Council.
  • Prakash, Abishur. 2019. “The Geopolitics of Artificial Intelligence.” Scientific American Blog Network (blog). July 11. https://blogs.scientificamerican.com/observations/the-geopolitics-of-artificial-intelligence/.
  • Rosenbrock, H. H., ed. 1989. Designing Human-Centred Technology: A Cross-Disciplinary Project in Computer-Aided Manufacturing. The Springer Series on Artificial Intelligence and Society. London ; New York: Springer-Verlag.
  • Ross, Casey. 2018. “IBM’s Watson Recommended ‘Unsafe and Incorrect’ Cancer Treatments.” STAT. July 25. https://www.statnews.com/2018/07/25/ibm-watson-recommended-unsafe-incorrect-treatments/.
  • Rowe, Gene, and Lynn J. Frewer. 2005. “A Typology of Public Engagement Mechanisms.” Science, Technology & Human Values 30 (2): 251–290. https://doi.org/https://doi.org/10.1177/0162243904271724.
  • Russell, Stuart J. 2019. Human Compatible: Artificial Intelligence and the Problem of Control. New York, NY: Viking.
  • Russell, Stuart J., and Peter Norvig. 2010. Artificial Intelligence: A Modern Approach. 3rd ed. Prentice Hall Series in Artificial Intelligence. Upper Saddle River, NJ: Prentice Hall.
  • Scharre, Paul. 2018. Army of None: Autonomous Weapons and the Future of War. 1st ed. New York ; London: W. W. Norton & Company.
  • Schmidt, Eric, Robert O. Work, Safra Catz, Steve Chien, Mignon Clyburn, Christopher Darby, Kenneth Ford, et al. 2020. NSCAI First Quarter Recommendations. Washington, DC: National Security Commission on Artificial Intelligence.
  • Schwab, Klaus. 2017. The Fourth Industrial Revolution. New York, NY: Crown Business.
  • Searle, John R. 1980. “Minds, Brains, and Programs.” Behavioral and Brain Sciences 3 (3): 417–457.
  • Searle, John R. 1999. “I Married a Computer.” The New York Review of Books, April 8. https://www.nybooks.com/articles/1999/04/08/i-married-a-computer/.
  • Searle, John R. 2014. “What Your Computer Can’t Know.” New York Review of Books, October.
  • Searle, John R., and Ray Kurzweil. 1999. “‘I Married a Computer’: An Exchange.” The New York Review of Books, May 20. http://www.nybooks.com/articles/1999/05/20/i-married-a-computer-an-exchange/.
  • Shoham, Yoav, Raymond Perrault, Erik Brynjolfsson, and Jack Clark. 2017. AI Index 2017. Stanford, CA: One Hundred Year Study on AI at Stanford University. http://aiindex.org/2017/.
  • Shoham, Yoav, Raymond Perrault, Erik Brynjolfsson, Jack Clark, James Manyika, Juan Carlos Niebles, Terah Lyons, John Etchemendy, Barbara Grosz, and Zoe Bauer. 2018. The AI Index 2018 Annual Report. Stanford, CA: Stanford University.
  • Silberman, Charles. 1966. The Myths of Automation. New York, NY: Harper & Row.
  • Smith, Elizabeth S. 1993. “On the Shoulders of Giants: From Boole to Shannon to Taube: The Origins and Development of Computerized Information from the Mid-19th Century to the Present.” Information Technology and Libraries 12 (2): 217–226.
  • Smith, Brian Cantwell. 2019. The Promise of Artificial Intelligence: Reckoning and Judgment. Cambridge, MA: The MIT Press.
  • Solo, Robert A. 1963. “Automation: Technique, Mystique, Critique.” The Journal of Business 36 (2): 166–178.
  • Stilgoe, Jack, Simon J. Lock, and James Wilsdon. 2014. “Why Should We Promote Public Engagement with Science?” Public Understanding of Science 23 (1): 4–15. https://doi.org/https://doi.org/10.1177/0963662513518154.
  • Strickland, E. 2019. “IBM Watson, Heal Thyself: How IBM Overpromised and Underdelivered on AI Health Care.” IEEE Spectrum 56 (4): 24–31. https://doi.org/https://doi.org/10.1109/MSPEC.2019.8678513.
  • Suchman, Lucy A. 1987. Plans and Situated Actions: The Problem of Human-Machine Communication. Cambridge, UK: Cambridge University Press.
  • Suchman, Lucy A. 2007. “Feminist STS and the Sciences of the Artificial.” In The Handbook of Science and Technology Studies, edited by Edward J. Hackett, Olga Amsterdamska, Michael Lynch, and Judy Wajcman, 3rd ed., 139–163. Cambridge, MA: MIT Press : Published in cooperation with the Society for the Social Studies of Science.
  • Susskind, Jamie. 2018. Future Politics: Living Together in a World Transformed by Tech. 1st ed. Oxford ; New York, NY: Oxford University Press.
  • Taplin, Jonathan T. 2018. Move Fast and Break Things: How Facebook, Google, and Amazon Cornered Culture and Undermined Democracy. New York, NY: Little, Brown and Company.
  • Taube, Mortimer. 1961. Computers and Common Sense: The Myth of Thinking Machines. New York, NY: Columbia University Press.
  • Tegmark, Max. 2017. Life 3.0: Being Human in the Age of Artificial Intelligence. 1st ed. New York, NY: Alfred A. Knopf.
  • Terborgh, George. 1965. The Automation Hysteria. New York, NY: W. W. Norton & Company, Inc.
  • Topol, Eric J. 2019. Deep Medicine: How Artificial Intelligence Can Make Healthcare Human Again. 1st ed. New York, NY: Basic Books.
  • United States Congress Senate Committee on Labor and Public Welfare. 1963. Nation’s Manpower Revolution: Hearings Before the Subcommittee on Employment and Manpower of the Committee on Labor and Public Welfare, United States Senate, Eighty-Eighth Congress, First Session, Relating to the Training and Utilization of the Manpower Sources of the Nation … U.S. Washington, DC: Government Printing Office.
  • Vaidhyanathan, Siva. 2018. Antisocial Media: How Facebook Disconnects US and Undermines Democracy. New York, NY: Oxford University Press.
  • Wachter-Boettcher, Sara. 2017. Technically Wrong: Sexist Apps, Biased Algorithms, and Other Threats of Toxic Tech. New York, NY: W. W. Norton & Company.
  • Webb, Amy. 2019. The Big Nine: How the Tech Titans and Their Thinking Machines Could Warp Humanity. New York, NY: PublicAffairs.
  • Webster, Graham, Rogier Creemers, Paul Triolo, and Elsa Kania. 2017. “Full Translation: China’s ‘New Generation Artificial Intelligence Development Plan’ (2017).” New America. https://www.newamerica.org/cybersecurity-initiative/digichina/blog/full-translation-chinas-new-generation-artificial-intelligence-development-plan-2017/.
  • Weinberg, Alvin M. 1962. “Review: Computers and Common Sense: The Myth of Thinking Machines by Mortimer Taube.” The Library Quarterly 32 (4): 309–310.
  • Weizenbaum, Joseph. 1976. Computer Power and Human Reason: From Judgment to Calculation. San Francisco, CA: W.H. Freeman.
  • Wheeler, Harvey. 1972. “Artificial Reasoning Machines and Politics.” In Computers and the Problems of Society, edited by Harold Sackman, and Harold Borko, 467–493. Montvale, NJ: AFIPS Press.
  • White, Lynn. 1980. Medieval Technology and Social Change. Reprinted. London: Oxford Univ. Press.
  • The White House. 2018. “Artificial Intelligence for the American People.” The White House (blog). May 10. https://www.whitehouse.gov/briefings-statements/artificial-intelligence-american-people/.
  • Wiener, Norbert. 1960. “Some Moral and Technical Consequences of Automation.” Science 131 (3410): 1355–1358.
  • Wiener, Norbert. 1964. God and Golem, Inc.: A Comment on Certain Points Where Cybernetics Impinges on Religion. Cambridge, MA: MIT Press.
  • Wiener, Norbert. 1989. The Human Use of Human Beings: Cybernetics and Society. London: Free Association.
  • Winner, Langdon. 1989. Autonomous Technology: Technics-out-of-Control as a Theme in Political Thought. Seventh Printing. Cambridge, MA: MIT Press.
  • Winograd, Terry, and Fernando Flores. 1987. Understanding Computers and Cognition: A New Foundation for Design. Reading, MA: Addison-Wesley.
  • World Economic Forum. 2016. The Future of Jobs: Employment, Skills and Workforce Strategy for the Fourth Industrial Revolution. Global Challenge Insight Report. Geneva: World Economic Forum.
  • Wynne, Brian. 2006. “Public Engagement as a Means of Restoring Public Trust in Science – Hitting the Notes, but Missing the Music?” Community Genetics 9 (3): 211–220. https://doi.org/https://doi.org/10.1159/000092659.
  • Yamagishi, Toshio. 2011. Trust. The Science of the Mind. Tokyo: Springer Tokyo. http://link.springer.com/https://doi.org/10.1007/978-4-431-53936-0.
  • Yampolskiy, Roman V. 2015. Artificial Superintelligence: A Futuristic Approach. Boca Raton, FL: CRC Press.
  • Zuboff, Shoshana. 2018. The Age of Surveillance Capitalism: The Fight for a Human Future at the New Frontier of Power. 1st ed. New York, NY: PublicAffairs.

Reprints and Corporate Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

To request a reprint or corporate permissions for this article, please click on the relevant link below:

Academic Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

Obtain permissions instantly via Rightslink by clicking on the button below:

If you are unable to obtain permissions via Rightslink, please complete and submit this Permissions form. For more information, please visit our Permissions help page.