20,401
Views
1
CrossRef citations to date
0
Altmetric
Original Article

Balancing privacy rights and surveillance analytics: a decision process guide

ORCID Icon, ORCID Icon & ORCID Icon
Pages 155-170 | Received 11 Jul 2020, Accepted 14 Apr 2021, Published online: 06 May 2021

ABSTRACT

The right to privacy has been discussed by scholars in multiple disciplines, yet privacy issues are increasing due to technological advances and lower costs for organisations to adopt smart surveillance. Given the potential for misuse, it seems prudent for stakeholders to critically evaluate Surveillance Analytics (SA) innovations. To assist in balancing the issues arising from SA adoption and the implications for privacy, we review key terms and ethical frameworks. Further, we prescribe a two-by-two Surveillance, Privacy, and Ethical Decision (SPED) Process Guide. SPED recommends the use of one or more of three ethical frameworks, Consequence, Duty, and Virtue. The vertical axis in the SPED matrix is the sophistication of an organisation’s SA and the horizontal axis is an assessment of the current privacy level and the rights afforded to the target(s) of surveillance. The proposed decision process guide can assist senior managers and technologists in making decisions about adopting SA.

1. Introduction

For many people, surveillance has a negative connotation. Surveillance is about control and power (Foucault, Citation1991; Marx, Citation2005b; Ragnedda, Citation2011; Vagle, Citation2016), but it is not inherently about coercion, discipline, covert spying, and loss of freedom. Surveillance poses a threat to privacy. In recent years, computer-based surveillance systems have been enhanced with powerful capabilities. For example, if we compare the state of the art in 2005 for video surveillance systems (Valera & Velastin, Citation2005) with current distributed, high-definition video monitoring hardware, and sophisticated software for facial recognition and real-time analytics (Garvie & Moy, Citation2019), we see how much has changed. Also, since 2000, the Internet and rapid diffusion of personal computing devices has created large volumes of structured and unstructured data that managers want analysed (Syed et al., Citation2013).

To meet the need to analyse opportunistic and intentional surveillance data, organisations have been hiring data scientists, business analysts, data engineers, marketing data analysts, and intelligence analysts (Power & Heavin, Citation2017). These data professionals are trained in various disciplines including Business Analytics and Data Analytics. Surveillance analytics (SA) involves both expert use of analytics tools including algorithms and statistics, and creation and deployment of embedded, “smart” surveillance through automated object detection methods leveraging high-complexity image/data processing technologies and algorithms (Hu & Ni, Citation2017).

Analytics and surveillance innovations are not always beneficial nor always harmful (Degli Esposti, Citation2014); hence, stakeholders, including key decision-makers such as senior managers, technologists, and data professionals must balance outcomes and make value judgements about adopting innovative analytics and surveillance. Making expanded use of business analytics (BA) with data about people and adopting SA innovations creates challenging ethical questions associated with privacy, data ownership, and accuracy (Moor, Citation2005).

People fear surveillance because of scenarios like George Orwell’s (Citation1949) “Big Brother” (Power, Citation2016) and articles on the Chinese surveillance state (Campbell, Citation2019). “Technology’s complexity and its potential to impact society have generated a certain ambiguity towards, sometimes even a fear of technology” (Palm & Ove Hansson, Citation2006, p. 547). Zuboff (Citation2015) metaphorically labels surveillance capitalism as “Big Other”. Many organisations are using SA including non-profits like hospitals and universities, and more traditional firms in banking, financial services, and professional services. Privacy problems are growing because of technological advances and lower costs for organisations to adopt smart surveillance.

This article explores issues associated with Surveillance Analytics (SA) in multiple settings and it attempts to provide guidance for professionals, especially those in business and data analytics, who want to ensure that a proposed surveillance innovation such as analytics will not unnecessarily infringe on privacy rights. The primary focus is on innovative surveillance adoption decisions in private sector organisations. This focus on private surveillance narrows the scope and complexity of the analysis to fewer types of settings with decision processes that are more understandable, and yet the number of specific decisions is likely much larger than in the public sector.

The next section defines and explores the key concepts of privacy rights, surveillance analytics, ethical decision-making, and balance (Section 2). We then frame both the need for and the challenge of balancing privacy rights in the context of surveillance analytics (Section 3). Section 4 examines some surveillance analytics innovations. Section 5 summarises our findings by proposing a prescriptive decision process guide. The SPED Process Guide recommends ethical frameworks for assessing surveillance analytics innovations and for evaluating the associated privacy implications. Section 6 discusses the agency problem and other associated issues. The concluding section draws some conclusions about next steps for both academics and practitioners.

2. Defining key concepts

Developing a decision guide to help those obligated to evaluate innovative surveillance analytics begins with defining and analysing key concepts and terms. This section defines the four concepts of balance, privacy and privacy rights, surveillance analytics, and ethical decision-making.

2.1. Balance and balancing

The goal of balancing the value of privacy rights that might be lost with any benefits from designing and deploying innovative surveillance capabilities implies a goal of finding a win-win resolution of conflicting outcomes. A balanced analysis examines what privacy rights must be diminished or reduced to balance the perceived benefits from adopting a specific surveillance analytics solution. Striving for balance means that there are no absolute privacy rights and that some innovations in surveillance create significant social good that cannot be ignored. According to Sableman (Citation2014), “Privacy policy, however, is inherently about balance. Almost every aspect of modern civilization interferes with personal privacy to some extent”. In 2020, it seems especially important for stakeholders to find a balance prior to implementing tools like video surveillance (Wolfe & SDMmag.com, Citation2020).

The balance metaphor does not imply that attaining the benefits from an innovation in surveillance analytics must lead to a decrease in privacy or a loss of privacy rights. Surveillance analytics is not in direct conflict with privacy rights. Cultural factors, trust, and prior privacy violations impact the perceived balance and perceptions of those individuals who are the targets of surveillance (Luppicini & So, Citation2016). Moosavian (Citation2016) does caution us that the balance metaphor can lead to misuse of confidential information. He notes benefits from a security analytics innovation are only subjectively commensurable with any harm from privacy infringement or reduction.

2.2. Privacy and privacy rights

The two related concepts of privacy and privacy rights are variously defined and are difficult to measure. Privacy rights especially depend upon laws in the jurisdiction where surveillance occurs. A common dictionary definition of privacy is “freedom from unauthorized intrusion and the quality or state of being apart from company or observation” (Merriam Webster, Citation2020). The International Association of Privacy Professionals (IAPP, Citation2020) define privacy as “the right to be let alone, or freedom from interference or intrusion. Information privacy is the right to have some control over how your personal information is collected and used”. Others suggest that privacy is one’s ability to control information about oneself (Bélanger et al., Citation2002; Stone et al., Citation1983). Clarke (Citation1999) states that “privacy is often thought of as a moral right or a legal right” (p. 60). Privacy of an individual is multi-faceted and includes (i) privacy of the person; (ii) privacy of personal behaviour; (iii) privacy of personal communication; (iv) privacy of personal data; (v) privacy of location and space; and (vi) privacy of thoughts and feelings (Borton et al., Citation2013; Lee et al., Citation2013). Privacy also refers to being able to keep certain acts and information, especially personal matters, to oneself and avoid public attention (Derlega & Chaikin, Citation1977).

Privacy rights refers to the concept that one’s personal information is protected from public scrutiny (P. M. Schwartz, Citation1994). In a foundational article, Warren et al. (Citation1890) reviewed the common law, the U.S. Constitution, and statutes related to a right to privacy. Then and now, privacy rights are not fixed, and they vary among jurisdictions. For example, the majority in Olmstead v. United States (Citation1928) ruled that incriminating evidence obtained in wiretapping by government officials did not violate constitutional rights. In his dissent to Olmstead, Justice Brandeis noted “time works changes, brings into existence new conditions and purposes”. Olmstead was overturned 40 years later in Katz and United States (Citationn.d.). Katz held that warrants were in fact required to wiretap pay phones. Also, Katz established a new test of whether there is an expectation of privacy in a situation upon which one may justifiably rely.

The United Nations Declaration of Human Rights Article 12 states that “No one shall be subjected to arbitrary interference with his privacy, family, home or correspondence, nor to attacks upon his honour and reputation. Everyone has the right to the protection of the law against such interference or attacks”. Some of the terms and phrases in Article 12 like “arbitrary interference” and “attacks upon his honour” are vague and subject to various interpretations.

2.3. Surveillance analytics

Surveillance analytics refers to the use of software algorithms to detect, classify, monitor, and track objects or persons in real-time or after an event (Zalud, Citation2013). Surveillance analytics is a process that uses analytics techniques and tools like statistics and machine learning to identify patterns in data collected about people. Surveillance systems monitor behaviour, activities, or other changing things to manage, direct, or protect people (Zuboff, Citation2015). Surveillance also involves intercepting private emails or phone calls, passive or active video camera data gathering and analysis, clicking behaviour on Facebook or YouTube, or monitoring web surfing.

Heibutzki (Citation2018) defines surveillance narrowly as “the covert observation of people, places and vehicles, which law enforcement agencies and private detectives use to investigate allegations of illegal behaviour. These techniques range from physical observation to the electronic monitoring of conversations”. Surveillance equipment includes various devices including audio recorders, digital cameras, GPS tracking devices, and real-time listening devices. Surveillance analytics applies business analytics (BA) methods and tools to data about people. BA is about the discovery of meaningful patterns in data (Delen & Ram, Citation2018). Also, as Power et al. (Citation2018) explain BA “applies qualitative, quantitative, and statistical computational tools and methods to analyze data, gain insights, inform, and support decision-making (p. 51)”.

New forms of social surveillance apply “scrutiny through the use of technical means to extract or create personal or group data, whether from individuals or contexts” (Marx, Citation2005a). Marx (Citation2005a) cites new tools like “computer matching, profiling and data mining; work, computer and electronic location monitoring; DNA analysis; drug tests; brain scans for lie detection; various self-administered tests and thermal and other forms of imaging to reveal what is behind walls and enclosures”.

Technologies such as knowledge mining and deduction, pattern recognition and cloud computing are widely utilised in the next generation of video surveillance systems (Xu et al., Citation2016). Surveillance creates “Big Data” and proposes a framework for how to process, organise, manage, and store massive video data (Xu et al., Citation2016). The term “Big Data” is widely used in both the academic and popular literature (Power & Heavin, Citation2017) and video surveillance data are the largest source of “Big Data” (Gandomi & Haider, Citation2015). According to Constantiou and Kallinikos (Citation2015), much of what comes under the “Big Data” umbrella is not collected deliberately and purposefully. Typically, “Big Data” is diverse, randomly collected and, not infrequently, trivial, messy, and agnostic (Anderson, Citation2008).

Big data is one outcome of surveillance analytics, as the use of video and other technologies results in the generation of large volumes of diverse data. These data may be used to monitor and track peoples’ behaviour as well as provide new opportunities to generate new data-based insights (Xu et al., Citation2016). However, Tene et al. (Citation2013) state “We live in an age of ‘big data’ (p. 239)”, but they note “the extraordinary benefits of Big Data are tempered by concerns over privacy and data protection” (p. 241).

2.4. Privacy calculus, privacy paradox, and communication privacy management

To explain individual reactions to surveillance, Culnan and Armstrong (Citation1999) propose the “privacy calculus” theory which seeks to explain how individuals rationally balance the benefits and costs of disclosing personal data (i.e., cost-benefit trade-off). Since then, the concept of privacy calculus has been extended (cf. Dienlin & Metzger, Citation2016; Dinev et al., Citation2008; Plangger & Montecchi, Citation2020). One of the reasons for this is reflected in the fact that individuals disclose personal information even when they consider that the risks are high (commonly referred to as “privacy paradox” (Barnes & State University of New York at Buffalo, Citation2020; Taddicken, Citation2014)).

Privacy paradox in the era of surveillance analytics is well documented (Doty, Citation2020; Rowe, Citation2020). In their work, Mourey and Waldman (Citation2020) provide insights in “which one’s subjective importance of privacy itself varies as a function of who is in control of managing privacy and the extent to which managing privacy is perceived to be easy or difficult” (p. 162). Another reason for extending the privacy calculus theory can be explained by individuals who have rule-based boundaries that they use to determine whether to conceal or disclose information from others due to the ubiquitous nature of technology and the level of its use by others (commonly referred to as ‘Communication Privacy Management – (Petronio, Citation2020)). As a result, it is important to explore the perceived sophistication of surveillance technologies (complexity) and the privacy level of control (certainty) over personal information when focusing on surveillance analytics as a single perspective will only provide limited results.

Surveillance systems are not inherently evil nor good, but they produce big data. However, due to a lack of transparency around the design, development, and use of surveillance data, citizens have privacy concerns. Across various cultures, privacy is considered a basic human right (Hartman, Citation2001). As a result, infringing on individual’s privacy unnecessarily is considered unethical (Hagen et al., Citation2018).

2.5. Ethical decision making

Ethical decision-making (EDM) is a process of evaluating and choosing among alternatives in a manner consistent with ethical principles and ethical frameworks (M. S. Schwartz, Citation2016). Ethical decision-making should be broader than applying lexical rules. In decision-making, lexical rules give one type of consideration absolute priority over another, e.g., privacy rights have an absolute priority and surveillance and surveillance analytics should never infringe or violate privacy rights, cf., Baron (Citation1986).

EDM is about “choosing the right thing to do”, it is rarely about clear-cut right and wrong decisions (M. S. Schwartz, Citation2016). In many situations, EDM is about choosing the “lesser of two evils” and finding a balance between principles and frameworks. Some technology decisions require a prioritisation among competing ethical values and principles. In most situations, there is a tendency to rationalise. Just because everyone is doing it or it is legal and permissible, does it mean that a new surveillance innovation is ethical? Ethics refers to applying “moral rules, codes, or principles which provide guidelines for right and truthful behavior in specific situations” (Lewis, Citation1985, p. 382).

Marx (Citation1998) argued that before implementing surveillance innovations, we should evaluate the proposed methods by asking 29 questions to help determine the ethics of a proposed surveillance innovation. For example, Marx suggests we ask, “Is the personal information used for the reasons offered for its collection and for which consent may have been given and does the data stay with the original collector, or does it migrate elsewhere?” Data professionals cannot be expected to memorise these questions, but checklists, processes, and frameworks can be an established component of EDM.

highlights the various EDM approaches (consequence, duty, and virtue) which have been reported in the surveillance analytics literature. This table emphasises the siloed approach to exploring EDM in surveillance analytics, with limited studies focusing on using multiple philosophical frameworks for ethical analyses.

Table 1. Overview of EDM in surveillance analytics studies

The next section frames the privacy and surveillance balance problem under investigation as part of this article.

3. Framing the privacy and surveillance balance problem

Surveillance and analytics are transforming nation-states and organisations into less personal environments (Zuboff, Citation2015) with greater centralised control. In many countries and organisations, people are directly and indirectly impacted by surveillance. Some uses of surveillance technologies and associated analytics have had significant benefits including creating new products, for example, in healthcare (Raghupathi & Raghupathi, Citation2014) and providing evidence about crimes whereas other uses like storing web cookies and gathering or harvesting Facebook users’ personal data have caused privacy concerns (Sewell & Barker, Citation2006).

The central problem addressed in this article is how to help decision-makers make balanced, ethical decisions about adopting a specific surveillance innovation. This frame helps us understand, define, and prioritise a complex, situational problem. The privacy-surveillance problem, as we have noted, is increasingly important and is often a decentralised decision made in organisations with broader societal implications (Zuboff, Citation2015).

Privacy and surveillance are not new topics. In 1986, Mason noted “Our society is truly an information society, our time an information age. The question before us now is whether the kind of society being created is the one, we want. It is a question that should especially concern those of us in the MIS community for we are in the forefront of creating this new society”. In his article, he focused on four ethical issues that pose a threat to human dignity: 1) privacy, 2) accuracy, 3) property, and 4) accessibility (Mason, Citation1986). All four of these issues remain relevant.

This analysis focuses primarily on adoptions of innovative surveillance technologies by organisations rather than political entities. Many organisations collect data for marketing purposes or data about employees. The hospital setting might focus on patient and employee data. A school or university may collect data on multiple groups including students, faculty, staff, donors, and alumni.

As technology and analytical capabilities evolve, we revisit topics. For example, the debate over monitoring Internet use and the email of employees in the workplace is ongoing. Alge (Citation2001) noted electronic workplace surveillance is raising concerns about privacy and fairness. In 2002, Zimmerman summarised the issues regarding privacy and monitoring. Little has changed in the debate other than arguing about new technologies, including the implementation of Video Monitoring Surveillance capabilities (SHRM, Citation2019).

At the time, Zimmerman and Workforce.com (Citation2002) was sceptical of IT involvement in surveillance decision-making. She quoted an IT consultant who equated IT staff to law enforcement. Supposedly he said, “you have the cops making the laws, and that’s not good”. Zimmerman and Workforce.com (Citation2002) also identified a reoccurring problem “that companies often make a snap decision about how they are going to use monitoring software”. Today there seems to be an understanding that both HR and IT experts should be part of an ethical decision-making process about when and how to monitor employees. As Dobrin (Citation2012) explains “Ethical problems are often complicated and require more than a formula to solve. The proper resolution of ethical problems requires judgment and good decision-making”.

For many years, researchers have discussed the growing surveillance threat. For example, Boyd and Crawford (Citation2012) asserted that “There is a deep government and industrial drive toward gathering and extracting maximal value from data, be it information that will lead to more targeted advertising, product design, traffic planning, or criminal policing (p. 675)”. The “Big Brother” metaphor mentioned in the introduction focuses our attention on data collection by government institutions in general and at all levels. Zuboff’s (Citation2015) “Big Other” metaphor expanded the threat to privacy from surveillance to larger corporations and other institutions. She asserts that any organisation that has sufficient resources to own and operate surveillance tools and maintain extensive data collections may invade privacy boundaries (Zuboff, Citation2015).

According to Richards and Harv. L. Rev (Citation2013), some possible harms from surveillance associated with diminished privacy rights include: 1) reduced intellectual privacy, for example, a chilling effect on discussing union organising, criticising management, or an inclination to conformity, 2) surveillance poses harm by altering the power dynamic between the watcher (the boss) and the watched (the worker or customer), and 3) surveillance may lead to blackmail and other corrupt behaviours. Richards and Harv. L. Rev (Citation2013) further notes if “we are watched while engaging in intellectual activities, broadly defined – thinking, reading, web surfing, or private communication – we are deterred from engaging in thoughts or deeds that others might find deviant. Surveillance thus menaces our society’s foundational commitments to intellectual diversity and eccentric individuality”.

Surveillance capabilities continue to expand with technological developments. Data professionals, including business analytics practitioners and data scientists are not trained as moral philosophers nor as lawyers and yet these experts increasingly need to make ethical decisions about new surveillance opportunities.

Some sources prescribe using Privacy by Design (Cavoukian, Citation2011; Hustinx, Citation2010) to protect privacy rights. Privacy by Design (PbD) is a system engineering approach initially intended for three application domains: 1) IT systems; 2) accountable business practices; and 3) physical design and networked infrastructure. PbD ideas can also be applied to business analytics and surveillance analytics. Ethical decision-making deliberations also help balance competing claims and concerns.

4. Surveillance analytics innovations

In a specific surveillance implementation, multiple analytics tools can be used to develop a system. For example, machine learning algorithms can be used for classification and prediction innovations like predicting feelings from facial recognition, cheating during test taking, and shoplifting (Gates, Citation2011). Real-time monitoring of millions of video and voice feeds is impossible without AI (Artificial Intelligence) technologies, especially Machine Learning. Cognitive computing, a term used by IBM to describe the current wave of AI, means software could even interact with people being monitored (Sommer, Citation2017). Both video and voice data could be archived, edited, and reviewed. Imagine organisations using Video Surveillance with facial recognition to monitor and track the actions of workers. Also, Video Surveillance has been widely used in schools to prevent bullying, deter vandalism, monitor visitors, and maintain a record of evidence in the event of a crime. Imagine new capabilities with alerts to school officials. One vendor advertises “Our HD Wi-Fi Nanny Camera Teddy Bear is the cuddliest guy around”. California, USA, is one of the few jurisdictions that requires the consent of all parties to a recorded conversation or video.

A broader, more ambiguous change that may threaten privacy involves the convergence of information and communication technologies. Surveillance analytics can exploit these advances by incorporating data sources from devices such as phones, video cameras, computer log files, specialised Internet of Things devices, and remote data capture devices (Zalud, Citation2013). Surveillance analytics algorithms and technologies leverage multiple data sources to provide descriptive, diagnostic, predictive, and prescriptive information. For example, Google provides cloud-computing services that include image identification, voice recognition, and machine learning technology.

Biometric technology like facial recognition systems can identify individuals without their knowledge. According to Sumner (Citation2016), in “November 2013 supermarket giant Tesco announced that it would be installing screens positioned by payment tills which scan its customers faces, then display targeted advertising to them”. One example of an innovative surveillance system is metaphorically called “Big Proctor”. Flaherty’s (Citation2020) article in Inside Higher Ed has received many comments. She explains “Online proctoring has surged during the coronavirus pandemic, and so too have concerns about the practice, in which students take exams under the watchful eyes (human or automated) of third-party programs” (Flaherty, Citation2020). She notes “Chief among faculty and student concerns are student privacy and increasing test anxiety via a sense of being surveilled” (Flaherty, Citation2020). According to an April 2020, Educause poll, 54% of institutions were using online or remote proctoring services, while another 23% were considering or planning to use them. Even so, over half of the institutions polled said they were concerned about cost, as well as student privacy. From a student or teacher perspective, ask yourself how you would assess an innovation like this with monitoring of head movements as an indicator of cheating. What data would you use to train a machine learning algorithm? Should you tell students about the algorithm? If the algorithm indicated a student cheated on a test, what would you do?

Rapidly improving technologies and remote distributed data gathering coupled with decision automation alters what is possible and what is expected by managers. With the continued changes in the application of analytics to big data involving new and more data sources and more sophisticated analytics techniques, some long-standing beliefs about fair data/information practices and provisions of existing law and guidance raise significant challenges for organisations that want to apply analytics to big data (Center for Policy Leadership, Citation2013).

The United Nations Office of the High Commissioner for Human Rights (Citation2020) website notes that “data-intensive technologies, such as artificial intelligence applications, contribute to creating a digital environment in which both States and business enterprises are increasingly able to track, analyse, predict, and even manipulate people’s behaviour to an unprecedented degree. These technological developments carry significant risks for human dignity, autonomy and privacy and the exercise of human rights in general, if applied without effective safeguards”.

In many political jurisdictions, people expect that surveillance data, especially any personal and behavioural information will be protected. In these jurisdictions, it is accepted that this information belongs to the person and does not belong to an organisation, the public, or government. Managers and stakeholders need to ensure that data that is collected and used does not infringe on the expected privacy rights of individuals (Power & Heavin, Citation2018). Regrettably, the exact extent of privacy rights for employees, customers, and other data providers is not always clearly defined.

Society is at an inflection point, and perhaps a turning point regarding the ethical use of surveillance analytics using advanced technologies. Identifying an ethical innovation in data collection and/or analytics applied to surveillance data requires answering a difficult and complex question.

5. A prescriptive decision guide

Adoption, appropriate use, and assessment of surveillance-related innovations should be made in the context of prevailing ethical standards (Breidbach & Maglio, Citation2020; McParland & Connolly, Citation2020). Surveillance and privacy perceptions provide a context for ethical decision-making about surveillance innovations (Darmody & Zwick, Citation2020). To assist in assessing and balancing surveillance data gathering and analysis proposals, we propose a two-by-two Surveillance, Privacy and Ethical Decision-Making (SPED) Process Guide that prescribes the most appropriate ethical decision-making process(s) given the decision context. According to Gregor (Citation2006 p. 620), a prescriptive approach specifies “how people can accomplish something in practice (e.g., construct an artifact or develop a strategy)”.

Using M. S. Schwartz’s (Citation2016) characterisation of types of EDM models, SPED is more interactional, i.e., person-situation focused, in nature. As a prescriptive framework, SPED “allows for the establishment of grounds (i.e., reasons) for imputation of a person with regard to her actions, of responsibility with regard to others, and of recognition with regard to unknown others” (Reijers & Coeckelbergh, Citation2020, p188). We argue that a single code of technology ethics cannot fit all contexts and situations.

The SPED framework draws upon principal-agent theory (Eisenhardt, Citation1989), also known as the agency dilemma, to provide an understanding of the forces that need to be examined in terms of balancing the privacy rights of individuals and the use of innovative surveillance innovations. Senior managers are the principals and employees or customers are agents. In a surveillance or monitoring situation, the principal, e.g., employer, assumes the role of the watcher and takes actions to protect the interests of various agents, e.g., employees who are watched. The implicit contract is often about creating a safe workplace for employees or providing customers with a good experience and value while maintaining their privacy rights. Agents may be concerned that the principal is acting from self-interest rather than their interest when implementing surveillance systems and using a data-driven approach in the form of surveillance analytics to predict behaviours.

The SPED framework (see ) can be viewed as a situation-specific extension of the macro-level contingency theory (Lawrence & Lorsch, Citation1967) to the micro-decision-making level. The two dimensions are framed specifically for privacy and surveillance analytics adoption decisions.

Figure 1. Surveillance, Privacy, and Ethical Decision-Making (SPED) Framework

Figure 1. Surveillance, Privacy, and Ethical Decision-Making (SPED) Framework

Taking Galbraith’s (Citation1974) information processing theory to the decision situation level, our guide prescribes that as uncertainty increases, the principal requires an increasing amount of information to make an ethical decision about the proposed surveillance analytics innovation. Duncan’s (Citation1972) complexity-dynamism hypothesis supports providing more information to the principal as complexity increases. In all four situations, one or more principal stakeholders act on behalf of a group of agents, e.g., employees or customers, who expect the principal to act in their best interests. SPED may shine a new light on the dilemmas, challenges, and information needs facing management decision-makers, data scientists, data and IT professionals, and government and organisation policy-makers.

The vertical axis in the SPED framework is a measure of the perceived sophistication of currently implemented surveillance analytics and technology. An organisation’s level of sophistication in the use and management of IT is a multidimensional construct with four components: 1) technology use and utilisation, 2) applications portfolio, 3) centrality of the IS function, and 4) managerial governance and control of IT (cf., Cheney and Dickson Citation1982); Raymond and Pare Citation1992).

Technology sophistication reflects the extent of diverse and complex IT elements employed within society (adapted from Coetzee, Citation2017) ranging from low to high on a continuum of sophistication. Low technology sophistication reflects simpler and generic technological elements. Therefore, decisions about the adoption of surveillance innovations in low technology sophistication situations can be informed by simple lexical rules and the application of the virtue ethical decision-making framework. In a high surveillance technology sophistication, situation with surveillance analytics in an organisation and/or country of operation stakeholders should make use of multiple ethical decision-making frameworks. Several types of ethical theory exist and using only one may not be sufficient to address ethical implications (Broad, Citation2014).

The horizontal axis in the decision guide is a measure of privacy levels and existing privacy rights of individuals in a country and a specific organisation. Privacy level is a measure of how much perceived control an individual has over how personal information is collected and used (Smith et al., Citation1996). Since 1997, Privacy InternationalFootnote1 has calculated an Internet Privacy index for countries. The index is calculated based on mean scores on 14 criteria, including Constitutional protection, Visual surveillance, Communication interception, and Workplace monitoring. When sophisticated surveillance already exists in a country or organisation, adopting more advanced surveillance analytics to process data creates a need for a more rigorous process for balancing the trade-offs of reduced privacy and additional benefits from surveillance (Anhalt-Depies et al., Citation2019; Pal et al., Citation2020).

Under conditions of low privacy, people have limited opportunities to identify which data, structured or unstructured, can be accessed, processed, and stored about them and thus, individuals are more vulnerable to unauthorised intrusion from surveillance. The opposite occurs in the high privacy situation as proposed by Tarkkanen and Harkke (Citation2016) when investigating privacy concerns in designing social network sites.

In , the concepts of surveillance sophistication and privacy level are identified as independent constructs, although the constructs are likely correlated with an inverse relationship where the lower the privacy level, the higher the surveillance sophistication. The relationship is most likely moderated by resource constraints and the availability of expertise due to the inequalities in digital proficiency experienced globally (Pearce & Rice, Citation2017). Also, small, and medium-sized organisations are much less likely to have sophisticated surveillance analytical systems. The constructs are represented as having only a high and a low value when the constructs are continuous variables, and the surveillance dimension is changing rapidly due to technological advances.

Both a current and ongoing assessment of surveillance technology sophistication and context factors are important in recommending an appropriate EDM decision process, as Morley et al. (Citation2020) argue “what was ethically justifiable in one place yesterday might not be so tomorrow”.

There are ethical implications for increasing surveillance of people in public places, but McParland and Connolly (Citation2020) argue that there may be even more issues when introduced in a workplace or within a person’s private domain. Hence, evaluating surveillance innovations is dependent upon context. There are three major components of the context that must be considered: 1) Place – where will the proposed surveillance occur? 2) Purpose – What is the purpose of the proposed surveillance? Perhaps a broad purpose like control or a more specific purpose like detecting changes in disease occurrence? and 3) Expectations of Privacy Rights – surveillance in one place may be ethical, but the same innovation may not be ethical in another place (Morley, Floridi et al., Citation2020). Likewise, surveillance for one purpose may be ethical but unethical for another (Koskela, Citation2006). Finally, in one political jurisdiction, privacy laws may permit or allow a specific type of surveillance, but the same surveillance may violate privacy rights in another jurisdiction (Gstrein, Citation2020).

Building from , the SPED framework contingently prescribes three normative approaches to ethical decision-making including Consequence, Teleological moral systems (Sinnott-Armstrong, Citation2019), Duty, Deontological moral systems (Alexander & Moore, Citation2016; Alston, Citation1988), and Virtue, Virtue-based ethical theories (Hursthouse, Citation1999) perspectives.

In the Consequence or Consequentialist framework, a stakeholder focuses on the future effects of the possible courses of action, considering the people who will be directly or indirectly affected. A decision-maker should ask what outcomes are desirable in each situation and consider ethical conduct to be whatever will achieve the best consequences (Howard-Snyder, Citation1996). A person using the Consequence framework desires to produce the greatest good (Driver, Citation2011). From this perspective, it is the consequences that make actions, good or bad, right, or wrong. The deliberative process should focus on outcomes and the likelihood they will be realised (Sinnott-Armstrong, Citation2019). The EDM process directs attention to the future effects of an action, for all people who will be directly or indirectly affected by the action. Ethical conduct is taking the action that will achieve the best consequences.

Alston (Citation1988) argues that the Duty perspective is one which follows rules or duty regardless of outcome. This approach harnesses obligations or duties of individuals and organisation to “do the right thing”. In the Duty framework, a stakeholder focuses on the duties and obligations the principal has in each situation and considers what ethical obligations exist and what one should never do (Larry & Moore, Citation2007). Ethical conduct is defined by doing one’s duties and doing the right thing, and the goal is performing the correct action. The deliberative process involves defining obligations in the situation, and what a decision-maker should never do (Lazar, Citation2017).

The Virtue framework relies on “virtuous traits” to guide ethical decision-making. This approach relies on human traits and behaviours as a way of discerning what is ethical. Virtue ethics states that practising good habits such as honesty and generosity makes a moral and virtuous person (Crossan et al., Citation2013). It guides a person without specific rules for resolving the ethical complexity. Crossan et al. (Citation2013) argues that the inclusion of a virtue ethical perspective into existing EDM frameworks is imperative for decision-making models. More recently, Drašček et al. (Citation2020) argue that virtue ethics is not considered in research on ethical decision-making. Luppicini and So (Citation2016) argue that ethical practices vary among individuals and communities. Hence, May (Citation2013) purports that virtue ethics represents a middle ground between duty and consequence ethics. Applying the virtue framework means a decision-maker focuses upon what a choice and my actions will show about my character. Ethical conduct is defined as what a virtuous (good) person would do in the circumstances.

Let us briefly examine the four quadrants in the SPED Framework in more detail:

Quadrant 1: Situations with high surveillance and analytics sophistication and with low levels of individual privacy create multiple ethical issues. To legitimise new surveillance projects proposed in Quadrant 1 situations, policy-makers, developers, and adopters of surveillance analytics tools and technologies and purchasers of surveillance-related data should use all three normative approaches to ethical decision-making, Consequence, Duty, and Virtue. Using these three perspectives together provides a more comprehensive and holistic view of the legitimacy issues associated with a surveillance proposal.

The three frameworks/perspectives are not mutually exclusive. The nature of ethics and ethical decision-making is so complex that these frameworks together may enable achieving a positive outcome.

Each of the three ethical decision-making approaches in this quadrant have potential weaknesses. From a Consequence perspective, successfully predicting outcomes is difficult to achieve, particularly when unexpected variables arise. There are also limitations associated with the Duty perspective, treating everyone the same may be ethical in principle but may result in a negative outcome for some individuals. Furthermore, with the Virtue perspective, it is difficult to define and agree upon what constitutes the term virtue and how it manifests in human behaviour as there are different virtue traits.

Quadrant 1 represents decision situations that currently have high surveillance and analytics sophistication and with low levels of individual privacy. In this situation, there is high complexity and low certainty so more information is easy to obtain.

Quadrant 2: Another scenario occurs when surveillance analytics tools and technologies have the capacity to capture a large volume of high velocity, and high variety structured and unstructured data about a specific person, but they have the option of identifying what data can be accessed, processed, and stored. The ethical decision-making process underpinning Quadrant 2 builds on Quadrant 1 but extends the decision-making scope to engage more individuals in the process and directs attention to the future effects of an action and to our duties. The legitimacy of surveillance is based on motivation to produce the best outcomes and to perform the right action. This decision-making approach, however, would be difficult to operationalise in real-time and is further complicated by the absence of approaches for identifying aberrant behaviour by adopters of surveillance analytics.

Quadrant 2 represents decision situations that currently have high surveillance and analytics sophistication with high levels of individual privacy. There is high complexity and high certainty, so more information is desirable.

Quadrant 3: Low amounts of surveillance and rudimentary surveillance analytics with low levels (limited) of individual privacy rights are perhaps the focus for the evolution of a “Big Brother” society. This quadrant attempts to discern character traits motivating people in the situation (i.e., Focus of Virtue Ethical Decision Perspective). Here, policy-makers, developers, and adopters of surveillance technologies and purchasers of surveillance-related data identify and examine a person’s initial impressions surrounding the capture, use and storage of personal data vis-à-vis surveillance technologies.

Quadrant 3 represents decision situations that currently have low surveillance and analytics sophistication and low levels (limited) of individual privacy rights. In this quadrant, there is low complexity and low certainty.

Quadrant 4: Surveillance technologies are limited in terms of data collection abilities and yet people have opportunities to identify what personal data can be accessed, analysed, and stored about them and thus people have a lower threat of unauthorised intrusion. This quadrant in the SPED Framework has the fewest ethical concerns because each person is in control of identifying and monitoring the capture, analytical processing, and storage of their personal data. Here the sole ethical responsibility is placed upon the individual (Virtue Ethical Decision-Making). As such, people will be faced with answering questions such as “What will my actions show about my character?” and/or “What would a virtuous person do in the circumstances?” Where people have high privacy rights, policymakers, and developers of surveillance technologies should provide an approach or system that enables a person to take control of their personal data and to determine how it is captured, processed, and stored.

Quadrant 4 represents decision situations that currently have low surveillance and analytics sophistication and high levels (limited) of individual privacy rights so people have a lower threat of unauthorised intrusion. Here, there is low complexity and high certainty.

Surveillance and analytics sophistication are positively associated with complexity. Privacy level is positively associated with the certainty of an individual about their privacy rights.

In a classic review article of ethical decision-making research, Ford and Richardson (Citation1994) examined variables associated with the individual decision-maker and variables which form and define the situation in which the individual makes decisions. They found some evidence of differences in behaviour based upon job responsibilities, but they concluded that “Industry ethical standards were not related to an individual’s ethical beliefs and decision-making behavior”. Prior research identified two situational factors, a positive organisational ethical climate, and the existence of corporate codes of conduct that positively increase an individual’s ethical beliefs and decision behaviour. They did note the type of ethical decision did influence decision-making behaviour. O’Fallon and Butterfield (Citation2005) and Craft (Citation2013) reported similar findings.

6. Discussion and reflections

Leveraging the premise of several theories, we propose the prescriptive Surveillance, Privacy, and Ethical Decision-Making (SPED) Process Guide to better support senior managers’ understanding of the forces that need to be examined to achieve “balance” when considering the privacy rights of individuals and the implementation and use of innovative surveillance systems. Technological advances have increased the possibilities for more sophisticated surveillance which can impact privacy, raising significant ethical questions. Our proposed framework acts to guide decision processes with the overarching goal of protecting the privacy rights of citizens. As a result, five recommendations are proposed to support people to develop a strategy which focuses on balancing Privacy Rights and Surveillance Analytics:

Recommendation 1 (based on Q1 SPED Framework): As this quadrant reflects high levels of complexity (high surveillance) and low certainty (low privacy), individuals will feel a loss of autonomy. Decision-makers should consider a range of ethical viewpoints to ensure that what, why, when, and how data is collected. Decision makers must consider the potential effect of a decision into the future, ensuring rules are adhered to but equally, to understand and appreciate rights and wrongs based on the consequences of these rules which are underpinned by reason and truth. Decision makers must build a reputation for consistently acting out their virtues over time. Herschel and Miori (Citation2017 p. 31) argue that a multiple philosophical ethical approach “affords insight into the context and the logic of the moral arguments being presented, thereby providing us with a rational mechanism by which to better evaluate whether an intended action or actual outcome is morally right or wrong”.

In doing so, this will provide insights for individuals to make more informed decisions about revealing and concealing private information (i.e., Communication Privacy Management). This theory states that individuals have rule-based boundaries that they use to determine whether to conceal or disclose information from others (Petronio, Citation2020).

In many ways, people, and especially information technologists and researchers, are caught in a paradoxical situation. Big Data and surveillance analytics algorithms and technologies have many positives for individuals, organisations, and society. For example, surveillance analytics are being used to understand individual citizen behaviour particularly in terms of compliance with “Stay at Home” guidelines and to track contacts in the event of a positive COVID-19 diagnosis. These technologies are being leveraged for the “greater good” of society, this strategy may be placed in Quadrant 1 of the SPED Framework where surveillance and analytics sophistication is high, and privacy rights are low resulting in serious ethical dilemmas.

Recommendation 2 (based on Q2 SPED Framework): As this quadrant reflects high levels of complexity (high surveillance) but high certainty (high privacy), the cost-benefit trade-off of disclosing personal data (Privacy calculus) will be observed by more individuals. To maximise this, decision-makers should embed a set of rules or principles which give full autonomy to individuals in managing how data is accessed, processed, and stored about them. Any future changes in the way surveillance analytics are employed by decision-makers should consider who might be affected and what the outcome of an action might be. Such an approach can be achieved through citizen-engagement activities and providing individuals with a dynamic informed consent process which enables them to opt in or out of certain surveillance data acquisition, use, and analysis. In this scenario, principal-agent theory may come into play where incentives or rewards are used to motivate the agent to provide consent. Arthur and Owen (Citation2019) present an example of this in financial services where customers can choose to opt into a scheme and consent to allowing their personal transaction data to be used in exchange for rewards provided under a merchant-funded rewards programme.

Surveillance analytics and privacy rights are local and organisational as well as personal and societal issues. The dynamics differ in each setting, but the ethical issues and the possibilities are similar. Ethical and responsible use of algorithms and surveillance data can have benefits (Stahl & Wright, Citation2018). With an increase in new types of data analysis techniques, there is a concern about a growing indifference to the specifics of persons, places, and events (Amoore, Citation2011, p. 30). Context matters in assessing the surveillance data so the place, time of day, and weather conditions where the data is originating among many factors must be considered in processing data, especially streaming data. This reality reinforces the need to ensure that surveillance technologies are developed, managed, and evaluated in a fair, transparent, explainable, and intentional manner (European Parliamentary Research Service, Citation2020).

Recommendation 3 (based on Q3 SPED Framework): This quadrant reflects low levels of complexity (low surveillance) and low certainty (low privacy). Since surveillance analytics is not human and privacy is commonly seen as an instrumental value in relation to personal autonomy, decision-makers who employ surveillance analytics should reflect upon their actions and assess if they are adhering to or ignoring the rights of others by their actions. One of the most important things to do at the beginning of an ethical deliberation is to identify ethical aspects of the proposed innovation and determine who may be affected by those making decisions. Van Der Sloot (Citation2014) argue that the virtue ethics embraced by decision-makers will enhance the background, value, and scope of the right to individual privacy, which he argues is “not only of theoretical importance; it has practical significance for privacy protection in the age of Big Data”.

In a human surveillance system, in the past managers watched employee and customer behaviours and drew conclusions based upon limited data and in many cases anecdotal experiences. Managers now can get better data and have experts and algorithms analyse the data for patterns and correlations. Decision-making can be informed and potentially more systematic. Managers and other data and privacy experts in an organisation must find the appropriate balance between the benefits of proposed surveillance innovations and the surveillance target’s expectations for privacy. For example, emails sent or received through an organisation email account are generally not private. Once a valid business purpose exists for the surveillance, employers are free to monitor and capture these communications using a proxy server. If the communications are analysed, the data analyst should assess the need to protect the identities of sender and receiver(s). An algorithm automatically assures anonymity and de-identification during analysis. A policy of de-identification can increase trust with those who provide personal information. Such a policy stated in the Terms of Service might be appropriate for data associated with store affinity cards, buyer rebate software, and discount or bank cards. Garfinkel and Department of Commerce, NISTIR Citation8053 (2015) asserts “De-identification can reduce the privacy risk associated with collecting, processing, archiving, distributing, or publishing information. De-identification thus attempts to balance the contradictory goals of using and sharing personal information while protecting privacy, p. iii”.

Recommendation based on Q4 SPED Framework: As this quadrant reflects low levels of complexity (low surveillance) and high certainty (high privacy), the decision-maker should always place the privacy control in the hands of the individual being monitored. The aim is to find a balance between monitoring and trust by building trust around the use of emerging technologies such as surveillance analytics (Bilal et al., Citation2020). It is argued that virtue ethics could be one solution to the privacy paradox (Bilal et al., Citation2020).

However, in this scenario, principal-agent problems may arise when information asymmetry exists. In the context of the pandemic where working from home or remote working has become essential, employers have had less oversight of employees’ daily activities. Some organisations have adopted a virtue lens for monitoring employee productivity with minimal surveillance. Holt et al. (Citation2017) outlines an alternative approach with organisations using monitoring systems to improve performance measurement, increase productivity, and reduce costs. In his Guardian article, Walker (Citation2021) outlines how one organisation plans to address information asymmetry with the implementation of home webcams connected to AI-based scanning systems to monitor the activity of 380,000 call centre staff working from home across 34 countries. The organisation has since refuted this claim stating that they trust their employees and that the innovative technology is being implemented to promote employee collaboration and to monitor customer data security. Trade unions and the UK government have communicated their concerns about the implementation of these surveillance technologies in the home.

Holt et al. (Citation2017) purport that the ethical implications of monitoring have been largely ignored as well as the impact on employees’ morale and their views of the organisation. There is an opportunity to place emphasis on virtues, or moral character, rather than duties, rules, or the consequences of actions (Bilal et al., Citation2020).

Recommendation 5 (Entire SPED Framework): Decision makers should be proactive in supporting the development and implementation of regulations that govern the use of analytics and AI in society, organisations, homes, and in devices. Using AI is a large scale “real world” experiment. We, as scientists, have an obligation to ensure that people are not harmed by algorithms and AI. All of us are becoming subjects in uncontrolled surveillance analytics experiments (Anderson & Rainie, Citation2018).

Some nations are moving, perhaps unknowingly, closer to the “Big Brother” society described by Orwell. Also, large multinational organisations often determine the balance of data control and economic power (Nemitz, Citation2018). On the one hand, individuals, national public health systems, and governments prioritise heath and access to the right healthcare above everything else. On the other hand, some privacy experts are beginning to flag the need for governments to implement appropriate data privacy and security measures now, advocating for “Responsible Surveillance” (Minevich, M. & Beridze, I., Citation2020).

7. Conclusions

This article explores issues associated with surveillance in diverse settings and it attempts to provide guidance for professionals in multiple disciplines, especially business and data analytics, who want to ensure that a proposed surveillance innovation such as analytics will not unnecessarily infringe on privacy rights. This article contributes to both theory and practice.

In terms of practical contributions, we propose a prescriptive two-by-two Surveillance, Privacy, and Ethical Decision-Making (SPED) Process Guide that can be used to guide the decision process in assessing ethical dilemmas arising from the use of surveillance analytics and their implications for privacy. The vertical axis in the SPED framework is a measure of the perceived sophistication of currently implemented surveillance analytics and technology. The horizontal axis is a measure of privacy levels and existing privacy rights of individuals in a country and a specific organisation. Stakeholders must want to make ethical decision for the framework to have value. Surveillance analytics and privacy rights seem to come into focus when new use cases or technologies become “visible” and highlight a potential problem.

It is problematic to justify actions as ethical simply because the data are accessible (Boyd & Crawford, Citation2012, p. 672). Business and data analysts and data scientists should critically review proposed uses of surveillance analytics for behavioural prediction. These specialist IT professionals should be the first line of defence in maintaining privacy rights that might be lost to surveillance innovation. Organisations should conduct ethics training for their IT professionals and have ethics committees in place to review the use of analytics with biometric data (cf., Forbes Insights, Citation2018).

From a theoretical perspective, we argue that a single code of technology ethics cannot fit all contexts and situations, leveraging a range of existing ethics and IS theories we present in the SPED Process Guide. This framework offers a theoretical lens for researchers to explore privacy rights through Consequence, Duty, and Virtue ethics against a landscape of increasingly sophisticated surveillance analytics. Future studies should include empirical research to evaluate the SPED framework and the proposed recommendations. Future research could consider how to leverage the SPED framework, 1) as a tool to support decision makers in their adoption and use of smart surveillance systems, and 2) as a tool to explore existing situations where surveillance technologies and big datasets have raised ethical concerns.

In her work on Big Other and Surveillance Capitalism, Zuboff (Citation2015) calls scholars and citizens to action “The trajectory of this narrative depends in no small measure on the scholars drawn to this frontier project and the citizens who act in the knowledge that deception-induced ignorance is no social contract, and freedom from uncertainty is no freedom (p86)”. Building on this, we assert that there is a real need to offer theoretical and practical guidance to academics and decision makers to enable and support the balancing of surveillance and privacy rights. This may prompt organisations to invest in and promote awareness of new training opportunities, guidelines, and policies to promote greater “balance”. The SPED Process Guide for decision-making is likely to apply in a variety of settings and levels (i.e., individual, organisational, governmental, and societal) where balancing the value of privacy rights and the use of innovative surveillance capabilities should be examined prior to the design and deployment of these solutions.

Acknowledgement

This article is dedicated to the memory of Prof. Daniel J. Power.

Disclosure statement

No potential conflict of interest was reported by the authors.

Notes

References