374
Views
0
CrossRef citations to date
0
Altmetric
Research Articles

Industry 5.0 use cases development framework

, , , ORCID Icon, &
Pages 6064-6089 | Received 08 Mar 2023, Accepted 09 Jan 2024, Published online: 31 Jan 2024

Abstract

The Industry 5.0 concept has placed human needs at the heart of industrial processes. This raises the question of how new technologies can enhance employee decision-making processes and influence the evolution of team autonomy. Recent studies have shown that the best way to measure these impacts is to conduct experiments in complex and realistic environmental settings. However, the main methods cannot satisfy this requirement while controlling the events and associated variables, whereas a set of use cases can. Therefore, a model should be defined to generate and structure these use cases while validating their relevance. Following the decomposition of the global research objective and case-definition recommendations, this study proposes a framework for designing complementary use cases to evaluate the impact of new technologies on emerging autonomy models in a structured, realistic, and global manner. Based on widely recognised related work, the 6-step framework helps define a coherent context specifying the business process model, agent, autonomy, technologies to be implemented, their fields of action, detailed variable collection protocol, and experimental setup. A cross-analysis of existing cases from the literature and empirical use of the framework validated the relevance of the model in designing experimental environments that are close to real-world settings.

1. Introduction

The term Industry 5.0, adopted by the European Commission (European Commission. Directorate General for Research and Innovation Citation2021), has emerged as a concept complementary to Industry 4.0. Various research and technology organisations and funding agencies agreed on the need to better integrate the EU's social and environmental priorities into technological innovation by shifting from individual technology to a systemic perspective. While Industry 4.0 places new technologies at the centre of production and supply chains (Roblek, Meško, and Krapež Citation2016), Industry 5.0 aims to reinforce this digital transformation through more meaningful and effective collaboration between humans, machines, and systems within their digital ecosystem. The partnership between humans and intelligent machines combines the precision and speed of industrial automation with human creativity, innovation, and critical thinking. With Industry 5.0, value-driven and human-centred initiatives overlay the technological transformations of Industry 4.0, creating more fluid interactions between humans and machines (Maddikunta et al. Citation2021; Müller Citation2020). In this study, the term Industry 5.0 was used. Over the last decade, organisations have focused on implementing new technologies to increase productivity, sometimes neglecting the human dimension (Eslami et al. Citation2021). In this context, the question arises as to whether these technologies at the interface between workers and industrial processes enhance workers’ autonomy in decision-making processes. The reality of industries shows that new technologies lead to a complete rethinking of their uses, which has the potential to radically transform work. To better measure this effect, further studies are required to capture the interactions between technologies (Xu, Xu, and Li Citation2018) and all stages of the decision-making process (Ivanov Citation2022). A safe and inclusive working environment is essential for prioritising autonomy, which is considered a worker’s fundamental right (European Commission. Directorate General for Research and Innovation Citation2021; Nahavandi Citation2019; Xu, Xu, and Li Citation2018). Notably, many experts and observers believe that the key element of I5.0 lies in placing humans at the centre of the decision-making process by collaborating with machines. As human-machine interaction develops within I5.0, autonomy appears to be a necessary condition for industrial resilience. Humans are developing a degree of dependence on systems, which they should overcome in case of a disruption. Moreover, the evolution of organisations prioritises autonomy (e.g. Lean Management, agile management, and frugal innovation). Consequently, we should ensure that the development of digitalisation does not undermine the emergence of this autonomy. Finally, autonomy is seen as one of the most fundamental aspects of the transition from a technocentric to a value-centric 5.0 industry (Enang, Bashiri, and Jarvis Citation2023). Thus, autonomy in human decision-making, supported by new technologies is particularly crucial for the future (Kumar et al. Citation2021).

In this context, research should focus on how new technologies improve decision making and impact employee autonomy. Such research is challenging, because studying real-world situations requires the evaluation or measurement of human-centric experimental variables that are complex and difficult to isolate. Fortunately, studying human-centred processes and technology transfer issues is possible in an observational laboratory (de Paula Ferreira et al. Citation2022; Zeisel Citation2020); however, it requires the development of appropriate scenarios. While most research is based on case studies (Nguyen Ngoc, Lasa, and Iriarte Citation2022), it does not permit a quantitative measurement of the performance and behaviour of the actors involved in a process. They have also been criticised for their subjectivity, data interpretation bias, and lack of rigour in their reproduction. Contrarily, use cases, which are artificial replications of real contexts, allow for experimentation in the laboratory and control of specific variables without anticipating the phenomena that will occur (Yin Citation1981b; Citation1981a; Citation2018). A use case is an empirical method that examines a contemporary phenomenon in depth and in its actual context, particularly when the boundaries between the phenomenon and context may not be clear (Yin Citation1981b; Citation1981a; Citation2018). The essence of a use case is to describe and explain the complex phenomena that occur in real life. A set of use cases built in a complementary and mutually consistent manner around an overall research question, can therefore, make it possible to study phenomena occurring in complex contexts similar to a case study, but without falling into their usual shortcomings. All these types of complex cases focus on research questions such as ‘how’ and ‘why,’ but without controlling the behavioural and contemporary events (Yin Citation1981b; Citation1981a; Citation2018). The model proposed in this study addresses this problem by decomposing an overall research question into specific, mutually consistent research sub-questions, resulting in the creation of a coherent set of use cases. This type of case has also been shown to be relevant when the main object of the study concerns decision-making, as it may be used to explain why decisions were made, how they were implemented, and what results were obtained (Meyer Citation2001; Schramm Citation1971). Use cases are primarily used in system engineering or user experience (UX) research to assess and clarify user behaviour while using technology or systems. Compared to a case study, a use case can help clarify the outcomes related to technology and its utilisation by controlling the effects of the utilisation context (Jacobson, Spence and Bittner Citation2011). This seems more coherent with the claim of I5.0 to improve and adapt the technology to the requirements of employees.

This study proposes a use case development framework to evaluate the impact of new technologies on new autonomy models in a structured, realistic, and comprehensive manner. This framework is closely related to our current work, which focuses on how enhancing the decision-making process through new technologies contributes to the emergence of new autonomy models for work centres. Beyond the scope of this study, this was our overall research question. This comprehensive research involves studying many complex phenomena in an environment as realistically as possible and requires consideration of a large number of variables of interest. Because the overall research question is extremely broad to be addressed in a single use case, we should create conditions for decomposing this overall research question into specific, mutually consistent, and specific research sub-questions, which can then be the subject of particular use cases. The framework proposed in this study is necessary for this decomposition.

The remainder of this study is organised as follows. In Section 2, we present a literature review of use cases that exploit new technologies. In Section 3, we outline the methodology used to develop our framework. In Section 4, we propose a framework for designing the use cases. Section 5 demonstrates the framework’s completion by analyzing the I5.0 cases reported in the literature. Section 6 discusses the results of the study, including the proposed framework and its validation. Finally, Section 7 concludes the study with a presentation of future perspectives offered by this framework and future research opportunities.

2. Literature review

2.1. People and technologies in Industry 5.0

Many authors have indicated that there is currently no clear definition of autonomy as applied to a production system (Everaere Citation2007). The term ‘autonomy’ comes from the Greek autos (oneself) and nomos (law, rule, organisation), and therefore refers to the idea of determining one's own rules or having the ability to govern oneself based on one's own rules (de Terssac Citation2012). Similarly, Cirillo et al. (Citation2021) considered that "autonomy is an expression of the leader's power, because it allows him to modify all actions". Brey (Citation1999) argued that worker autonomy is related to ‘the control that workers have over their own work situation,’ and draws on a definition of job autonomy as ‘the worker’s self-determination, discretion or freedom inherent in the job, to determine several task elements’ (De Jonge Citation1995). These task elements include the method of work, pace of work, procedures, scheduling, work criteria, work goals, workplace, work evaluation, working hours, work type, and amount of work. Brey (Citation1999) argued that worker autonomy refers to the degree to which employees have control over some or all task elements. Moreover, it is difficult to imagine that autonomy has no impact on collaboration within the team in which the operator works. The role of I5.0 employees will evolve toward that of decision-makers actively involved in a decision-making process that considers the whole context (Frazzon et al. Citation2013; Schuh et al. Citation2014). In this study, the notion of autonomy will be considered in the context of these visions. Ultimately, it refers to the freedom an agent has to make a decision, which leads us to define the concept of decision making. Simon (Citation1960) is among the first scholars to formulate a decision-making model. According to him, a decision begins with an investigation phase that involves identifying the gap between the current and desired situations. This is followed by a design phase to define possible actions to resolve the situation, which were then compared and selected in the final selection phase. In this study, we adhered to this definition and restricted ourselves to decisions arising from a situation gap analysis. The agent concept used here is similar to that established by Macal and North (Citation2009): an autonomous decision-making entity that receives sensor information from an environment and acts based on that information. Finally, this research was conducted on the scale of the work centre, which APICS defines as ‘a specific production area, consisting of one or more people and/or machines with similar capabilities, that can be considered as one unit for capacity requirements planning and detailed scheduling’ (Pittman and Atwater Citation2022).

European Commission (European Commission. Directorate General for Research and Innovation Citation2021) identified Industry 5.0, as a complementary and expanded vision of Industry 4.0, which recognises the importance of creating employee-centric factories to be more resilient and sustainable. Simultaneously, operational work is affected by technology. Technology can also empower employees to achieve higher productivity, as demonstrated by MIT’s Work of the Future Initiative (Autor et al. Citation2022). Assessing the implications of new technology adoption on employees is imperative, particularly when health, learning, performance, and decision making are affected (Pinzone et al. Citation2020). Meindl et al. (Citation2021) showed that the interfaces among the people involved, technologies used, and operational processes for improving work are not often explicit.

Recent studies in the field of operation management have investigated aspects of the dynamics between new technologies and work, addressing topics such as new relationships between people and technology (Longo, Nicoletti, and Padovano Citation2017; Peruzzini, Grandi, and Pellicciari Citation2020), the impact on work design (Cagliano et al. Citation2019), and the adoption of wearables in industrial settings (Maltseva Reiby Citation2020; Zheng, Glock, and Grosse Citation2022). Fifteen operation-related technologies were identified in a literature review by Dornelles et al. (Citation2022). For example, augmented and virtual realities are often mobilised to help operators perform complex tasks faster and with significant confidence (Uva et al. Citation2018). These are used in assembly operations (Lai et al. Citation2020), training (Tao et al. Citation2019), and quality control (Szajna and Kostrzewski Citation2022). In these assembly operations, augmented reality is used to indicate the assembly steps to follow (De Pace et al. Citation2020), signal and prevent errors, and communicate to the employee remotely with the supervisor or engineer in case of doubt (Calzavara et al. Citation2020).

Of note is the emergence of technologies that operators can use in various work situations. These wearables are generally used to collect data on operator movements, concentration, and state to improve working conditions and ergonomics (Guo et al. Citation2019; Sun et al. Citation2020). Wearables are used in assembly for the real-time collection of operator data, mainly on movements and workflows (Maltseva Reiby Citation2020). However, no research has been conducted on the use of these technologies to verify their impact on engagement or decision-making.

Complementary studies have highlighted the contributions of technologies to production planning, control (Bueno, Godinho Filho, and Frank Citation2020), and decision-making activities (Ivanov Citation2022). These mechanisms have been extensively studied at the strategic level (Olhager and Feldmann Citation2022). However, the mechanisms of decision-making at the operational level are yet to be explored in the context of implementing new technologies (Ivanov Citation2022). The latter also illuminates the impact of new technologies on employee decision-making autonomy (Rosin et al. Citation2021).

2.2. Use cases in operations management research

Research on I5.0 is mostly based on literature reviews (Panagou, Neumann, and Fruggiero Citation2023), text mining (Grosse Citation2023), interview analyses (van Oudenhoven et al. Citation2022), numerical experiments (Abdous et al. Citation2022), and industrial case studies (Kaasinen et al. Citation2020). The use of rich experimental settings in operation management research is limited. Gao, Li, and Sun (Citation2022) identified only 192 experiment-based publications, the majority of which were based on non-field experiments, such as laboratory or quasi-experiments. Whether they are field experiments, in organisations, or in controlled laboratory settings, they provide a rich means of establishing causal links between the multiple elements under study (Eden Citation2017) and lead to a finer understanding of the mechanisms under study and behaviours of agents (Highhouse Citation2009).

Laboratory experiments can be designed to test analytical models and verify the underlying theory by providing scenarios that are similar to real ones (Katok Citation2011). Researchers have also called for combining laboratory-type and experimental field approaches (Gao, Li, and Sun Citation2022) to increase the validity and generalizability of the research results.

Given the increasing centrality of technology and how employees use it, the development of use cases seems to fulfil two needs: to verify the impact of the technology on the user and consider the requirements and context of user. Use cases offer a unique advantage by providing operations close to real-world settings in controlled environments. In the literature, use cases are generally scenarios replicating specific organisational processes and technologies, which is important as there is often a dynamic interplay between the technology in its ecosystem (Maghazei, Lewis, and Netland Citation2022). Therefore, a call for use cases has been issued (Saihi, Awad, and Ben-Daya Citation2021). However, the few use cases that have been realised either focus on the technology primarily or on one type of technology, such as digital twins (Attaran and Celik Citation2023), IoT, and CPS (Lesch et al. Citation2023), leaving little room to interpret how this affects the individual in the workplace. Employee behaviour and ethical aspects of using technology seem to be missing from existing use cases (Ordieres-Meré, Gutierrez, and Villalba-Díez Citation2023). However, two recent studies have addressed the impact of technology on employees. The first presents the impact of a more or less powerful artificial intelligence (AI) on an operator's engagement, stress level, and cognitive load (Passalacqua et al. Citation2023), whereas the second enriches the same use case by adding augmented reality technology to the experimental protocol (Joblot et al. Citation2023). If research on this topic is to be conducted in the future, it will be essential to propose a framework for its structure.

When a use case approach is used, it tends to be simple (Katok Citation2011) or focused exclusively on a particular technology (Rožanec et al. Citation2022). The design of a use case is often missing. This can be partially explained by the lack of specific models for designing use cases to measure the impact of technology use on employees. Among the available approaches, Ordieres-Meré, Gutierrez, and Villalba-Díez (Citation2023) offered an architecture that integrates human and machine data to improve operational transparency. Golan, Cohen, and Singer (Citation2020) proposed a framework for investigating future operators (i.e. workstation interactions in the I4.0 era). A complex system can identify the degradation of an operator's performance or system state and correct it through different interventions. Moencks et al. (Citation2022) introduced a tool to guide practitioners in their decision-making processes. This tool takes a macro view of human-technology interfaces and ensures that their implementation generates value. It does not address the more concrete aspects of the direct impact of new technologies on humans, their autonomy, or their ability to make decisions. Autonomy is only considered when considering manufacturing systems (Mo et al. Citation2023). Moreover, none of these models allow for the design of larger use cases involving multiple actors, and measurement methods are not addressed.

By proposing a more structured approach to conducting these empirical studies through use cases, we aim to follow the trend of empirical approaches on the relationship between technology, decision-making, and work in Industry 5.0 (e.g. Dornelles et al. Citation2022; Peruzzini, Grandi, and Pellicciari Citation2020).

3. Methodology

This study aims to propose a framework for designing complementary use cases to assess the impact of new technologies on autonomy in an operational context in a structured, realistic, and comprehensive manner. This framework should make it possible to structure a set of use cases to answer the following global research question: ‘How does the enhancement of the decision-making process through new technologies contribute to the emergence of new autonomy models for work centres?’

3.1. Choice of empirical research method

Yin (Citation2018) compared different empirical research methods that can be applied to this type of study and suggested distinguishing them based on three conditions:

  • The nature of the research question,

  • The degree of control the researcher has over actual behavioural events, and

  • The extent to which this study focuses on contemporary events is contrary to that on historical events.

Because our research question is of an explanatory nature, framed as a ‘How’ question, and primarily focuses on contemporary events, two methods stand out: case studies and experiments. These two methods differ in terms of the control a researcher has over actual behavioural events:

  • Case studies are employed when the researcher has limited or no control over events and the boundary between a phenomenon and its context cannot be clearly delineated. Case studies allow for an in-depth examination of a contemporary phenomenon (referred to as the ‘case’) within its real-world context, which entails the consideration of numerous variables of interest.

  • Conversely, experiments require the researcher to exert direct, precise, and systematic control over actual behavioural events. The experiments aimed at isolating and focusing on the phenomenon of interest by deliberately separating it from its context. These tests are typically conducted in controlled laboratory settings; however, field experiments are also possible. Experiments concentrated on one or a few selected variables to establish causal relationships.

Both case studies and experiments have their strengths and are applicable to different research scenarios, depending on the research question and the level of control desired over the events being studied.

The generic research question addressed in our work calls for the study of complex phenomena that can only be understood after many variables of interest have been considered. This eliminates the need for further experiments. It also aims to study these phenomena in a realistic context such that the results are credible and exploitable in a real environment. However, the number of possible intersections between new technologies and the different ways in which they can be mobilised to reinforce a decision-making process implies that the number of contexts to be studied is extremely large. Therefore, case studies are unrealistic.

Literature frequently describes and analyzes the implementation of use cases to study the use of new technologies for I5.0. However, to the best of our knowledge, no generic methodology exists for this design. They represent a compromise between use cases and experiments for studying complex phenomena in a context that closely resembles a real environment, in which event control is not systematic.

3.2. General framework design methodology

Given the complexity of this phenomenon, we aimed to study the most realistic environment possible, and it was necessary to consider many variables of interest. Therefore, we employed a case study methodology to structure our framework for designing complementary use cases. The methodology proposed by Yin (Citation2018) serves as a widely recognised reference, which we used as a foundation.

Yin (Citation2018) indicated that in the context of a case study, a research plan is structured around five key components:

  • Case study questions;

  • Its proposals, if any;

  • Its case(s);

  • The logic linking the data to the proposals; and

  • Criteria for interpreting results

The design of the proposed framework is based on an approach that defines the components that constitute a use case and its scope (How? Why? What? Who? Where?):

3.2.1. Breaking down the defined global research question

Following the recommendations of Yin (Citation2018), the framework was designed to adhere to a logical sequence that connects the empirical data and conclusions to the research questions addressed in the use case.

Given that the overall research question is extremely broad to be addressed through a single-use case, it is necessary to create conditions for breaking down this research question into specific and mutually consistent sub-questions that can be addressed through individual use cases. Thus, the overall research questions were broken down as follows:

  • The research question is necessarily a ‘how?’ or ‘why?’ question.

  • The phenomenon under study: Enhancement of the decision-making process through new technologies (what?), and their potential impact on the emergence of new autonomy models (on what?).

  • The contextual elements impacted by the phenomenon. In the case of our global research question, these agents (who?) are likely to have their autonomy affected.

  • Other contextual elements to be characterised: the environment in which the phenomenon under study occurs (where?)

Note that at this stage, triggering the global research question, the context can only be partially defined; however, it should at least be expressed in the form of ‘How/why a phenomenon (what?) has an impact on something (on what?) that affects some of the contextual elements’.

3.2.2. Characterise the phenomenon under study, its impact, and its relationship

Defining the ‘what’ is central to the research plan. We characterised the phenomenon under study, its impact, and the relationship between the phenomenon and its impact. The phenomenon studied encompasses two key objectives: decision-making process and technology types. To characterise the decision-making process, we employed the model proposed by Rosin et al. (Citation2021), which builds on the model of Mintzberg, Raisinghani, and Théorêt (Citation1976). The same authors created a model outlining different autonomy types based on this decision-making process. We referred to the ten technology groups proposed by Danjou, Pellerin, and Rivest (Citation2017) to characterise the technology types. This classification draws upon and enriches the widely cited classification of Rüßmann et al. (Citation2015). To establish the relationships between the studied phenomenon and the object of its impact (i.e. new autonomy models), we formalised these relationships through the matrix structure presented in Subsection 4.5 (Step 5). This structure is based on the work of Rosin et al. (Citation2022), who examined the potential of new technologies to enhance the decision-making process and their connection to new autonomy types (Rosin et al. Citation2021).

From this characterisation of the phenomenon studied, the object of its impact, and its relationships, it is possible to identify a coherent set of specific research questions that can be learned through use cases. Each use case that can be extracted from the framework aims to answer a particular research question of the type ‘How does the reinforcement of specific steps in the decision-making process by a specific set of new technology groups contribute to the emergence of a specific autonomy model for operational teams?’

3.2.3. Characterise elements of the context

Two contextual elements should be distinguished: those directly affected by the phenomenon under study and other factors that contribute to defining the context of the use case.

First, we should characterise the agents (who) affected by the phenomenon and their relationship with the object of the phenomenon's impact, which in the context of our study is autonomy. For this analysis, we relied on the work-centre concept defined by APICS (Pittman and Atwater Citation2022). The autonomy referred to in this study can be achieved by one or more individuals who potentially interact with one or more machines.

Based on the characterisation of the studied phenomenon, the object of its impact, and their relationships, along with the specific research question at hand and the characterisation of the agents’ level of autonomy, we can formalise the propositions (as identified by Yin Citation2018) that we aim to validate through future use cases.

Finally, we should characterise other contextual elements that are not directly affected by the phenomena under study. In our case, they correspond to the organisational framework chosen for the study (where?). Designing use cases (cf. Yin's components) requires defining and bounding cases (Yin Citation2018). This involves determining the organisational framework (where) in which the phenomenon under study occurs. This is the subject of Step 1 in the framework. In our study, the latter generally relates to ‘the enhancement of decision-making processes using new technologies.’ Here, we should link the decision-making process model enhanced by new technologies to the operational process model within which decisions are made.

In Step 1 of our framework, the design of the organisational context is based on the definition of the business process model, which defines the tasks within business processes for which decision-making is required (Object Management Group Citation2023). Specifying the exact scope of the business process reproduced in the use case, which may be a subpart of a global business process, is essential. The organisational framework used to support use cases can occur in more or less complex environments, depending on whether the use cases are developed in a company or learning laboratories. At the end of this step, it is possible to properly define and formalise the cases to be studied, which correspond to one of the components identified by Yin (Citation2018).

3.2.4. Characterise the type of variables of interest and the type of measurements

In the first instance, we consider the cognitive, affective, and behavioural variables of interest to be characteristics of autonomy (Gagné and Deci Citation2005; Hackman and Oldham Citation1976; Oldham and Cummings Citation1996; Peruzzini, Grandi, and Pellicciari Citation2017). Concentration and mental absorption during task performance characterise the cognitive variables of interest. The affective variable of interest groups included perceptual aspects (reaction and mobilisation of the five senses) and emotional factors (valence of positive and/or negative emotions, as well as activation of emotions) during task performance. The behavioural variable of interest was characterised by observable elements of autonomy in task performance.

There are many different data collection methods in Industry 5.0 research, as summarised by Passalacqua et al. (Citation2022). Although the I5.0 research is still in its early stages, questionnaires are the most widely used data collection method. Psychophysiological methods have also been used, but neurophysiological methods have not. We took advantage of all types of data collection related to an experimental case (in the laboratory, in situ, or through use cases) to extend the scope of use of the framework, which we describe in more detail below (cf. Table ). Table  summarises the results of the literature analysis on the types of measurements. It refers to the main measurement methods used (column 1) (self-reported, observational, and physiological), in which articles (column 3), and simultaneously indicates the variables of interest they could measure in the articles in question (column 2).

Table 1. Measure types associated with examples of variables according to current literature.

The three variables of interest and the three identified measures constitute Step 6, encompassing the final components outlined by Yin (Citation2018). Additionally, while we focused on measuring the conditions under which autonomy emerges, it was crucial to ensure that these conditions do not compromise performance. Therefore, the performance was consistently measured in each use case to guarantee maintenance. The added value of the framework lies in providing a means to measure the perspective of an individual on what is transpiring.

Thus, the proposed framework encompasses six steps, collectively covering yin’s essential components. Furthermore, the performance measurement in each use case ensures that autonomy is studied without detriment to performance. The framework offers a valuable contribution by providing means of measuring the viewpoint of an individual.

4. Use cases development framework

The proposed framework aims to support different research methodologies by defining a series of use cases that lead to experimentation in an observation environment, allowing for qualitative and quantitative data collection without disrupting the actual functioning of an organisation.

The framework permits the detailed design of use cases in six steps. As shown above, the use case generated proposes the five elements defined by Yin (Citation2018) required for its correct definition: a case study question, its proposals, if any, its case(s), the logic linking the data to the proposals, and the criteria for interpreting the results. This framework is embedded in a more general research approach guided by the overall research question (‘how does the enhancement of the decision-making process through new technologies contribute to the emergence of new autonomy models for work centres?’). The framework enables this overall research question to be broken down into several specific research questions (‘how does the reinforcement of specific steps in the decision-making process by a specific set of new technology groups contribute to the emergence of a specific autonomy model for operational teams and systems?’) and provides use case(s) for each specific research question. Thus, the framework ensures a coherent approach to the overall research question in a structured and rigorous manner.

Step 1 defines the business process section to be studied using a use case to structure it around a concrete element. This step clearly defines the perimeter within which the agent evolves and exercises autonomy and decision making.

Step 2 defines the agent and its autonomy dimensions. This step ensures that human beings are involved in the use cases as early as possible.

Step 3 defines the autonomy granted to an agent. This step structures the decision-making possibilities of the agent and defines the full scope of its autonomy in the identified sections of the business process.

Step 4 defines the decision-making step(s) enhanced by new technologies. This step establishes the decision(s) that the agent is called upon to make in the use case.

Step 5 defines group(s) of new technologies to enhance previously selected decision-making step(s). It determines the proposals for the use case based on the perceived capacity of the new technologies selected in the context defined here.

Finally, Step 6 defines the experimental part of the use case aimed at testing the proposals described in the first five steps of the framework. This step consists of selecting the variable(s) of interest and measurement protocol(s) for these experiments, and structuring different experimental configurations.

The general framework is illustrated in Figure . The following subsections describe each step of the framework.

Figure 1. Industry 5.0 use cases development framework.

Global view on the six steps and all choices of the Industry 5.0 use cases development framework: 1. Select business process model; 2. Select agent characterisation; 3. Select autonomy type; 4. Select decision-making step; 5. Select group(s) of new technologies; 6. Select variables of interest and data collection protocol.
Figure 1. Industry 5.0 use cases development framework.

4.1. Step 1: select business process model (where?)

As a reminder, Step 1 defines the section of the business process to be studied using the use case to structure it around a more concrete element. This step clearly defines the perimeter within which the agent evolves and exercises autonomy and decision making.

An agent’s autonomy is expressed through well-defined business processes. However, the experiment linked to the use case does not involve the entire business process. Therefore, in the first step, the business process step (or step transition) should be precisely targeted. The business process may be taken from a real-life example of opportunism or meeting a specific industrial need. The aim here is to target the relevant step that requires the agent to make a decision. However, it may be useful to recreate a context adapted to the phenomenon that we wish to study. Thus, the business process can be considered complete. The first step is to define it completely before targeting the relevant step in the use case.

The definition of context then continues to Step 2, which characterises the agent.

4.2. Step 2: select agent characterisation (who?)

Step 2 defines the agent at the core of the use case. This agent is similar to all or part of a work centre, that is, one or more individuals, potentially interacting with one or more machines. It is the agent's autonomy in the face of the decisions they should make that the use case enables us to study. The agent can perform a set of tasks related to the business process defined in Step 1 and enjoys varying degrees of freedom in organising its work to accomplish them. This autonomy can have a strictly personal dimension (how an operator organises their tasks) or a more collaborative dimension (how operators interact to organise work-centre tasks). The highest level of organisational maturity enables complete collaboration between actors, objectives are shared, and work routinely follows coordinated and synchronised operations (Mo et al. Citation2023).

We defined an agent as a human entity observed and studied throughout the use case. The agent is qualified by one of the following two statuses: operator or team. An operator is a single individual or a machine that performs a specific technical operation. They are partly or wholly responsible for the performance and scope of actions (task execution and decision making). A team is a group of individuals and/or machines sharing a collective work situation, subject to common objectives and mutual responsibilities (Piquet Citation2009). Thus, responsibility is linked to each member’s actions and expected results.

Most importantly, this involves defining the dimension of autonomy from which the agent benefits. The focus can be on the autonomy specific to the task or, more globally, on its impact on collaboration. If the focus of autonomy is on the task, the use case relates to the autonomy of the agent in performing their tasks. This concerns the agent's authority and freedom to define their tasks: actions on the sequence of tasks, method of execution, pace of work, and tools used to perform the work. If the focus of autonomy is collaboration, the use case will focus on the agent's power to influence the organisational and collective environment: involvement in improving the work organisation, ability to influence decisions, and dynamics of cooperation at work. Therefore, collaboration between individuals and/or machines in the work centre can be studied, similar to collaboration between different work centres.

In this step, the agent is identified, and the focus is on the scope of his/her autonomy (on their tasks or collaboration within or outside the work centre). Thus, the agent should be the main actor in the business process defined in Step 1: The next step delimits the autonomy granted by the agent.

4.3. Step 3: select autonomy type (on what?)

Step 3 defines the autonomy granted to an agent. This step structures the decision-making possibilities of the agent and defines the full scope of its autonomy in the identified sections of the business process. This subsection is further divided into several sections: Based on Rosin et al. (Citation2021), the first section presents the decision-making model chosen for this framework. The second section (Section 4.3.1. to 4.3.7) presents the seven autonomy types derived from the model by Rosin et al. (Citation2021). Finally, to help the user, the last section (Section 4.3.8) proposes tools for classifying and identifying the autonomy type most likely to correspond to the use case that they wanted to design.

Different types of autonomy can be described or structured based on the decision-making processes proposed by Rosin et al. (Citation2021): Following Mintzberg, Raisinghani, and Théorêt’s (Citation1976) model, this process is divided into three phases: (1) problem or opportunity validation, (2) solution validation, and (3) implementation validation. Figure  illustrates this model.

Figure 2. Decision-making process (Rosin et al. Citation2021).

Process illustrating the steps followed by agents in making decisions: it starts with capture measure, followed by gap recognition, optionally followed by diagnosis, followed by search and then selection if known solutions or followed by design if unknown solutions, followed by evaluation and optionally followed by authorise.
Figure 2. Decision-making process (Rosin et al. Citation2021).

The problem/opportunity validation phase includes the Capture-Measure and Gap Recognition steps. The capture measure step collects real-time information from the production system, whereas the gap recognition step recognises an abnormal situation that requires a response. We then proceeded to the solution validation phase through Diagnosis, Search, Design, Selection, and Evaluation steps. The Diagnosis step aims to characterise the problem by establishing cause-and-effect relationships in the situation under study to determine whether solutions already exist and proceed to the search step, or whether the situation is new and should move to the Design step. The Search step is used to determine the solution(s) most likely to solve the problem. The Design step is used to design a new solution. If multiple solutions exist, this leads to a Selection step that acts as a filter to reject inappropriate solutions. Finally, the Evaluation step compares the solutions and validates whether the selected solution solves the problem. Then comes Phase 3 and its single step, Authorisation. Authorisation to implement the solution is provided by the operator, machine, or higher authority. This generic model allows for definition of the seven types of autonomy.

4.3.1. Type 1: cyber monitoring

In this type of autonomy, the cyber-physical production system (CPPS) should identify a situation or stimulus that triggers an analysis and decision. The decision-making process is then completed by the teams responsible for managing the situation without further assistance from the CPPS. Cyber monitoring scenarios include the Capture-Measure and Gap Recognition steps that generate stimuli that lead to a decision. By enabling more data to be captured and analyzed in real time, new technologies can immediately, or in some cases predictively, identify performance gaps, errors, and problems in production. The decision-making process can then be initiated more quickly to identify the actions to be taken, thereby improving the operational efficiency.

4.3.2. Type 2: cyber search

For this type of autonomy, the CPPS should propose one or more solutions to an encountered problem based on a pre-established set of possible corrective actions. Faced with an identified situation, the cyber search scenario reinforces the search and diagnosis steps to quickly analyze and target known solutions to correct a problem or respond to an opportunity. The attention and working memory of an agent are particularly challenged at this stage of the decision-making process, and are critical factors that limit the interpretation of information from the environment. Simulation and immersion logic can also reinforce the diagnosis step by allowing real time comparison of the current situation with the situation simulated on a virtual replica of the production system.

4.3.3. Type 3: standard decision support

In this autonomy type, the CPPS should identify a problem, identify a set of possible solutions, and, after possibly filtering them, evaluate the most relevant one(s) and propose a viable solution. The specificity of this scenario reinforces the Evaluation step of the decision-making process. This step is preceded by a Selection step, which provides one or more possible solutions. Based on systematized data processing, the Selection step aims to limit the number of solutions to be processed subsequently in the Evaluation step, which is generally more restrictive in terms of the time and complexity of implementation, as it aims to identify, among the selected solutions, the one likely to meet the set objectives. Previous research has shown that an agent recognised for its expertise in operational decision situations evaluates a plan of action using mental simulation to anticipate what would happen if this plan were applied in the context of the current situation (Klein Citation2008). Simulation and immersion technologies are vital to support operational teams and reduce the cognitive load required for this step.

4.3.4. Type 4: cyber control

This autonomy type goes beyond the standard decision support by reinforcing the Authorise step and facilitating the implementation of the action plan selected in the evaluation step by transmitting the necessary information to the operational level. This last point does not necessarily imply task automation because information can be passed on to the operational team for subsequent translation into action. The Authorise step is generally necessary when applying the chosen solution involving a scope of responsibility other than that of the production centre managing the problem or opportunity. This approval may require horizontal information sharing across organisations. This may be the case, for example, when the root cause of the encountered problem or a key lever for action lies outside the scope of the operation team. Authorisation may also require a vertical flow of information. This occurs when approval to implement the chosen solution is linked to other tactical or strategic decisions. New technologies that improve the horizontal or vertical integration of systems are required at this stage of decision-making.

4.3.5. Type 5: cyber design

The cyber design type is characterised by the reinforcement of the Design step to develop tailor-made solutions, either when the operational team should handle an unknown situation or when no known solution is perfectly adapted to the current situation. In an operational context, the activation of the Design step usually occurs after known solutions have been searched for and evaluated without success. In a completely unknown situation, the Design step can be initiated immediately after a problem or opportunity is identified. There are two prominent cases: the ‘pure design’ case, where tailor-made solutions should be developed without relying on already known solutions, and the case where solutions are modified from already known alternatives. In an operational context, the latter is preferred because it is generally less time consuming, costly, and demanding in terms of the resources and skill levels required. For this reason, it seems more interesting to consider implementing cyber search or standard decision support types beforehand such that the feedback loops leading to the design step can take place as soon as possible after the Search-Selection or Evaluation steps.

4.3.6. Type 6: customised decision support

A unique feature of customised decision support is the enhancement of the Evaluation step after a custom solution is developed in the Design step. This type is ultimately similar to the standard decision-support type in terms of enhanced steps; however, it has the distinction of evaluating previously unknown custom solutions. This modified context introduces subtleties to the evaluation process. Multicriteria decision methods are more appropriate for assessing standard solutions associated with prototypical situations. When used alone, they are not suitable for customised decision support. This type of autonomy first requires the implementation of simulation technologies aimed at simulating and numerically testing the evolution of the production system based on the newly envisaged solutions. Generally, this involves specifying an action plan associated with the implementation of a given solution. It is then possible to translate this action plan into parameters to be updated (which often corresponds to solutions developed by modifying an already known standard solution) or into a scenario to be tested in the simulation model. The simulation then supports decision analysis to assess each possible future state's likelihood and usefulness of future state, and estimate the maximum and minimum achievable results.

4.3.7. Type 7: cyber autonomy

This last autonomy type differs from the previous one in that it reinforces the Authorise step when needed and reinforces the implementation of the action plan for customised solutions. The Authorise step is generally used more often here than for the cyber control type, because implementing a customised solution is more often conditioned by validation from another area of responsibility than that entrusted to the operational team. Strengthening the Authorise step facilitates the implementation of delegation logic. This notion, already well known in organisations, consists of a manager with authority over the operational team, transferring part of their responsibilities and the ability to act and make decisions (Verrier and Bourgeois Citation2016). This is generally accompanied by control exercised by the manager, whose rules are easier to define in advance when standard solutions are already known.

4.3.8. Classification and identification of autonomy types

It should be noted that each autonomy type differs from the others in the nature of the steps in the decision-making process enhanced by new technologies and by the fact that the solution to be designed, validated, or authorised is already known. If we ignore the fact that the solution is known, then the types of autonomy can be classified into four main categories: enhanced data collection (Type 1), enhanced solution search (Types 2 and 5), enhanced solution evaluation (Types 3 and 6), and enhanced solution authorisation (Types 4 and 7). Similarly, the types of autonomy can be divided into two classifications: standardised solutions (Types 1, 2, 3, and 4) and customised solutions (Types 1, 5, 6, and 7). These two classifications and the four categories resulted in a matrix that maps the types of autonomy along the two axes, as shown in Figure .

Figure 3. Classification of autonomy types. Adapted from Rosin et al. (Citation2021).

The seven autonomy types classified in a matrix with four columns (enhanced data collection, enhanced solution search, enhanced solution search and evaluation and enhanced solution search, and evaluation and authorisation) and two rows (standard solutions and customised solutions).
Figure 3. Classification of autonomy types. Adapted from Rosin et al. (Citation2021).

To quickly identify the type of autonomy involved in the use case, the user should answer a series of questions, as illustrated in Figure .

Figure 4. Flowchart for autonomy type selection.

Flowchart that starts by choosing which stimuli trigger the decision-making process to then decide which type of autonomy to address by asking the questions of solution knowledge, authorisation, and multiplicity of solutions.
Figure 4. Flowchart for autonomy type selection.

What types of stimuli trigger decision-making? The stimulus may be related to deterioration, in which the current performance differs from that of the previous situation. In this case, the solution may already be known, leading to the classification of standardised solutions, or it may be a new solution, leading to a second classification of customised solutions. Alternatively, the stimulus may be an opportunity–that is, the possibility of achieving, under certain controlled conditions, a level of performance that is better than that of the current situation. This is generally outside the context of standardised work and its gap analysis is based on differences from a standard; the solution is much more likely to be completely new. This classification is more suitable for customised solutions.

Are all solutions known? When the solution is known, we automatically move toward the types of autonomy in the standardised solution classification, namely Types 2, 3, and 4. Conversely, when the solution is unknown, Types 5, 6, and 7 from the customised solution classification are favoured.

Is hierarchical approval required? When the approval of the hierarchy is needed and extended, the autonomy types will inevitably come from the enhanced solution authorisation classification.

Are there multiple solutions available? This question allows us to assess the usefulness of the evaluation step and thus, the extent to which the technologies invest in the decision-making model.

Once the autonomy type is selected, the next step is to identify the decision-making step(s) most relevant to the use case.

4.4. Step 4: select decision-making step(s) (on what?)

This step defines the decision-making step(s) enhanced by new technologies. This step defines the decision(s) that the agent is called upon to make in the use case.

As described previously, the decision-making process includes eight steps. For a specific type of autonomy, studying all the proposed steps can be counterproductive, and it seems wise to limit the number of steps to be studied. To achieve this, it is necessary to distinguish between the necessary, optional, and unnecessary steps in each model. Concurrently, some essential and optional steps will be enhanced by one or more new technologies, whereas others will still be present but will be performed without strong technological support. Figure  shows the necessary and enhanced technologies for each of the seven types of autonomy.

Figure 5. Enhanced decision-making step(s) in the chosen type of autonomy.

Table comparing all autonomy types pairwise which helps identify decision-making step(s) that are enhanced, non-enhanced, or unused.
Figure 5. Enhanced decision-making step(s) in the chosen type of autonomy.

This figure illustrates the similarities among certain types of autonomy. The most complex types of autonomy, including the maximum number of steps enhanced by new technologies, can be perceived as extensions of the less complex type of autonomy. The cyber autonomy type can be considered as the equivalent of a customised decision support type to which the Authorise step has been added and enhanced. Consequently, studying a complex type of autonomy is of interest only if the study focuses on steps specific to that type of autonomy. Thus, it seems inadequate to build a use case around the cyber autonomy type if the study focuses on the Capture-Measurement step. Contrarily, cyber monitoring type allows the same study without the added complexity of later steps.

At this stage, the context of the use case is well-defined. The specific research question then takes the following form: ‘How does the reinforcement of specific steps in the decision-making process contribute to the emergence of a specific autonomy model for operational teams and systems?’ To answer this specific research question, the next step is to select the group(s) of new technologies to be included in the use case.

4.5. Step 5: select group(s) of new technologies (what?)

This step defines group(s) of new technologies that enhance previously selected decision-making step(s). This step determines the use case proposals according to the perceived capacity of the new technologies chosen in the context defined here. This subsection introduces the concept and usage of a relevance matrix before explaining how to design it using an example adapted from Rosin et al. (Citation2022).

4.5.1. Concept and usage of a relevance matrix

This step is structured around a relevance matrix between the new technologies and eight decision-making steps. At each intersection of this matrix, there is a capability indicator that can be green, white, or red. The structure of this matrix is illustrated in Figure .

Figure 6. Relevance matrix between groups of new technologies and the eight decision-making steps.

Fill-in-the-blank table showing the eight decision-making steps in columns and the ten groups of new technologies in rows.
Figure 6. Relevance matrix between groups of new technologies and the eight decision-making steps.

The use case proposal(s) were directly derived from this capability indicator. The notions of the enhanced and control configurations are more clearly defined in Step 6.

A green indicator indicates positive consensus on the capabilities of the technology involved in enhancing the decision-making process. The proposal is therefore of the following form: [group(s) of new technologies] promote the autonomy of the [agent] in its decision-making by enhancing its ability to [decision-making step(s)] in [section of the business process]. The modus operandi is based on the agent’s autonomy and is more significant in the enhanced configuration than in the control configuration.

A red indicator shows a negative consensus. The technology does not appear to be suitable for enhancing the involved decision-making steps. Therefore, the proposal has the following form: [group(s) of new technologies] do not promote the autonomy of the [agent] in its decision-making because it does not enhance its ability to [decision-making step(s)] in [section of the business process]. The modus operandi is based on the fact that an agent's autonomy is just as significant, if not less significant, in the enhanced configuration as in the control configuration.

A white indicator indicates a lack of consensus on the impact of the technology concerned in enhancing the decision-making steps involved. The proposal is the same as that for a green indicator and is therefore of the form: [group(s) of new technologies] promote the autonomy of the [agent] in its decision-making by enhancing its ability to [decision-making step(s)] in [section of the business process]. The modus operandi has no particular expectations but aims to detect the slightest effect on the agent's autonomy in both augmented and control configurations.

4.5.2. Design of a relevance matrix

This matrix can be completed by various means: a literature review for a theoretical matrix, survey of experts for a more pragmatic matrix, and experiments for a more realistic matrix. By interviewing a Delphi–Régnier panel of equal numbers of academics, experienced industrialists and new technology providers, a list of new technologies most likely to support one or more of the eight steps of the decision-making process was generated (Rosin et al. Citation2022). This type of work fits perfectly into this step because it represents the current industrial and scientific requirements and remains valid in a broader operational context. The results of this study are presented in Figure  in the expected format at this step of the framework.

Figure 7. Proposed relevance matrix of new technologies to the steps in the decision-making process. Adapted from Rosin et al. (Citation2022).

Table mapping the capacity of the new technologies to support each decision-making step. Each case is displayed with red, white, or green colours to illustrate the capabilities of the technologies in the decision-making steps.
Figure 7. Proposed relevance matrix of new technologies to the steps in the decision-making process. Adapted from Rosin et al. (Citation2022).

These expert recommendations can be used to define correct combinations of scenarios and technologies. If certain technologies such as AI and the cloud seem to be naturally linked to several autonomy types, the study also shows experts’ dissent in applying some of these new technologies. Nevertheless, Figure  can guide researchers to define the right new technologies that the use case will explore and the type of experimentation the use case will be able to support.

Figure 8. Step 6: Observational variables and measuring methods.

Two circles side by side. One depicts the three variables of interest that can be studied (cognitive, affective, and behavioural), and the other shows the three types of measures (self-reported, physiological, and observational) that can be used to collect the variables of interest. This is integrated into a reference framework where performance measure is omnipresent.
Figure 8. Step 6: Observational variables and measuring methods.

At the end of this step, the specific research question is completed, proposal(s) are identified, and the context of the use case is established. All that remains is to define the logic linking the data to the proposals and criteria for interpreting the results in Step 6.

4.6. Step 6: select variables of interest and data collection protocol

This step defines the experimental part of the use case aimed at testing the proposals defined in the first five steps of the framework. This subsection is divided into two sections. The first section describes how to select the variable(s) of interest and measure the protocol(s) for the experiments. The second one explains how the structures of the different experimental configurations.

4.6.1. Variable of interest and measure

The final step in the framework is to identify the variables to be collected directly from the use case during the various experimental phases. Note that these data remain entirely independent of the status of the agent, although they seem to correspond to the capabilities of an individual. We are interested in both the capabilities of an individual and those of a team or an organisation. The last step is to select the correct variable(s) of interest to be measured in the use case. However, selecting one or more variables of interest is insufficient if a data collection protocol is not defined. The measurement protocol (measures) is as important as the type of data (variables) to be measured. In this step, the user should choose the variables of interest and the way to measure them. Figure  presents the three variables of interest and their measures.

The variables were divided into three main categories: cognitive, affective, and behavioural. Cognitive variables relate directly to an agent's mental abilities such as memory, attention, and decision-making. Affective variables are broadly related to sensory perception by combining visual acuity with olfactory sensitivity, intuition, and emotions. Finally, behavioural variables refer to the agent’s behaviour or variations in behaviour, such as initiative, excitement, stress, or aggressiveness.

Three types of measurements can be defined: self-reported, physiological, and observational. Self-reported measures are those in which participants report their thoughts, feelings, or behaviours. This was accomplished through surveys, questionnaires, and interviews. These measures are subject to bias, because the reported facts are first interpreted by the individuals reporting them. Physiological measures assess both physiological and bodily functions. These parameters include heart rate, blood pressure, brain activity, or hormone levels. These measures are much more objective, but usually require invasive equipment. Observational measures assess observable actions or behaviours. Examples include counting the number of times a behaviour occurs, determining the duration of a behaviour, and observing and rating the quality of the behaviour. These measures remain partially objective because observer bias can influence them.

Finally, Figure  shows that these variables of interest were measured in addition to the performance measures. It is crucial to ensure that the new technology group(s) do not degrade the agent's performance while performing its business process. Therefore, the agent's performance was continuously measured to ensure that it remained in the enhanced configuration, at least similar to that of the control configuration.

4.6.2. Enhanced configuration and control configuration

Several configurations were used to develop an experimental protocol in line with the use case proposals. In this study, we call ‘enhanced configuration’ any experimental configuration involving one or more groups of new technologies supporting the agent, and we call ‘control configuration’ the configuration without any new technology supporting the agent. Therefore, the principle of experimentation was to compare one or more enhanced configurations with a control configuration.

Certain new technologies and their utilisation can differ depending on their capabilities, that is, low, medium, and high. Thus, it is possible to design configurations that are low, medium, and/or highly enhanced. In addition to the comparison between these enhanced configurations and the control configuration, the chosen implementation conditions were used to verify the impact of the configuration modification over the course of the experiment.

  • Maintaining capability: no configuration change,

  • Increasing the capability of new technology: moving from a low-enhanced to a high-enhanced configuration

  • Degrading the capability of the new technology: moving from a highly enhanced to a low-enhanced configuration, or

  • Stopping/Adding capability: moving from an enhanced configuration to a control configuration and vice versa.

However, the control group remained in a control configuration throughout the experiment.

5. Validation

The validation of this framework is based on three elements:

  • Theoretical validation was conducted through a precise description of the methodology applied to the framework design (cf. Section 3 Methodology). This validation ensures that all framework design choices, as well as all tools referred to throughout the six steps, have their origins in the recognised work.

  • Empirical validation (below) presents an experiment: the step-by-step use of the framework to generate a relevant use case. This experiment ensured that the framework could generate a relevant use case containing all the information required for its experimental application.

  • Retrospective validation (below) compares the different choices offered by the framework at each step with data from actual cases as illustrated in the current literature. Finally, this validation ensures that the framework is complete or at least as exhaustively as possible.

5.1. Empirical validation

It is essential to ensure that the framework generates comprehensive and relevant use cases. To demonstrate this, we provided an experiment: the application of the framework to generate an actual use case. We adhered to the structure of the framework and provided a detailed account of the choices made at each step.

Step 1: The business process under consideration is that of the snowshoe assembly line in our factory laboratory. This step requires isolating the business process section that we wish to study. We then focused on the final assembly operation. There are two actions here: perform a quality control of the sole received and, if it is correct, finalise the assembly by inserting two latches. Otherwise, snowshoes are rejected. Therefore, we have a concrete decision-making situation (Is the sole correct?). The business process is clearly defined at the beginning and end, and the actions and decisions involved are identified.

Step 2: The business process, particularly the selected section, helps select the agent's characteristics. The decision we are interested in here is made by an agent with operator status, who is focused on their own tasks. Nevertheless, it would have been possible, for example, to focus on the decisions that the operator could make to ensure that an identified quality defect no longer occurs on the assembly line, thus integrating a cooperative dimension.

Step 3: We used the framework's classification of autonomy types (Figure ). In the context of our use case, and more specifically, in the section on the business process selected, it appears that the agent's decision derives directly from its ability to detect quality defects. Because gap analysis is at the core of cyber monitoring, we chose autonomy Type 1. However, we could imagine a business process in which the operator does not reject the defective sole but instead corrects it. Selecting autonomy Types 2 or 5 to enhance the solution search (known for Type 2 or unknown for Type 5) would then be more relevant.

Step 4: The table for this step in the framework (Figure ) implies that the capture measure and gap recognition decision-making steps are reinforced in autonomy Type 1. The simple discovery of a gap between the analyzed and standard sole is sufficient for the agent to decide to discard it. Therefore, we focus on these two decision-making steps and, more specifically, on the gap recognition decision-making step, which appears to trigger the decision. However, we could have imagined asking the agent to qualify for the gap and thus integrate the diagnostic decision-making step into the use case.

Step 5: We used the relevance matrix constructed from the Delphi–Régnier study by Rosin et al. (Figure ). The gap recognition decision-making step is thus highlighted, and experts agree that AI is relevant for enhancing this decision-making step. It would be interesting to propose a use case to validate this assertion within a specific context. The specific research question is complete: ‘How does the reinforcement of the gap recognition step in the decision-making process by AI contribute to the emergence of a cyber-monitoring autonomy type for operational teams and systems?’ Moreover, the resulting proposal would be: ‘AI promotes the operator autonomy in its decision-making during routine tasks by enhancing its ability to recognise a gap as part of a quality control in an assembly process.’ We can also select (or add) another group of new technologies such as augmented reality. In this case, because this group of new technologies caused dissensus among experts, the aim of the use case was to clarify this dissensus.

Step 6: In the final step, we define the measures and variables of interest and set up various experimental configurations for the first experiment. We aim to observe and measure the implications of the operator on the implemented Artificial Intelligence. We then selected the behavioural variables of interest. To limit any bias, we selected physiological measurements using equipment to capture heart rate, respiratory rate, sweating, etc. Next, we defined two experimental configurations: the control configuration featuring an operator without AI support and enhanced configuration featuring an operator with reliable AI support in 100% of the cases. During the first experiment, we made one configuration change; thus, the testing was stopped. The control group remained in the same configuration throughout the experiment. Given the positive consensus on AI in the context of our use case, it appears that the proposal will take a step towards confirmation if the agent's involvement is found to be greater in the enhanced configuration than in the control configuration. Finally, we added brain activity measurement equipment to extend the experiment and measure the cognitive load. We also defined a medium-enhanced configuration involving 80% reliable AI to further vary the experimental protocol.

This experiment, which illustrates each step of the framework, demonstrates that it is possible, in just a few steps, to define a coherent and relevant use case, offering a clear and precise context accompanied by several experimental protocols that would validate the proposals put forward, and thus answer a specific research question that fits automatically into a more global research approach. The works presented by Passalacqua et al. (Citation2023) and, more recently, by Joblot et al. (Citation2023) present use cases representative of the proposed example and incorporate all these elements.

5.2. Retrospective validation

However, the framework was compared with case studies and use cases found in the literature on I5.0. The restriction to I5.0 papers ensures that the selected cases are recent and deliberately human-centred. The literature on I5.0 includes several articles that use case studies and use cases to assess the impact of new technologies on employees. For a complete analysis of the literature, case studies were sufficiently close to use cases for inclusion in the search for articles for this validation. Experiments were conducted in realistic and complex environments. We sought to validate the relevance of our framework by analyzing whether it could propose use cases that mimic the characteristics of these cases.

The main aim of this validation was to ensure that the framework is complete, that is, all the parameters of the studied cases are proposed in one of the steps of the framework. The current literature was analyzed to validate the completeness of the framework.

The validation process started by identifying journal and conference articles in the Scopus database, exposing case studies and use cases focusing on Industry 5.0 practices. The query used was the following: TITLE-ABS-KEY ((‘industrie 5.0’ OR ‘industry 5.0’) AND (‘case stud*’ OR ‘use case’)) AND (LIMIT-TO (DOCTYPE, ‘cp’) OR LIMIT-TO (DOCTYPE, ‘ar’)) AND (EXCLUDE (SUBJAREA, ‘MEDI ORENER ORPHYS ORPSYC ORARTS ORMATE’)). The search returned 118 papers by December 2023.

Before using this set of articles to validate the framework, it was essential to ensure that each article encompassed a case study focused on assessing the influence of one or more groups of new technologies on an agent, preferably within a decision-making context. The initial reading of each article facilitated the exclusion of articles that primarily addressed technology implementation validation, training individuals on new technologies, or limiting accidents between humans and machines. Finally, 21 articles exposing relevant cases were retained and analyzed further. Articles written by (Rožanec et al. Citation2022) and Longo, Padovano, and Umbrello (Citation2020) propose multiple cases, which brought our total to 24 relevant case studies and uses cases.

The framework was validated using the Table  presented below, which serves as an analytical framework for each case identified from the literature. Each case is represented by one row in the table. The table is divided into five coloured sections corresponding to Steps 2–6 of the framework (Step 1 does not propose any choice). Each section was further divided into items that characterised the respective framework steps. Each coloured section includes two additional columns.

  • The column at the beginning of the section indicates whether a particular case can be characterised by any item within the considered framework step.

  • An ‘other’ column at the end of the section indicates if a given case was characterised by an item not explicitly structured within the framework.

Table 2. Validation of the use cases development framework.

The validation process involved completing a table by marking the corresponding boxes when an item (column) was identified within a specific case (row). The final row of the table serves as a summary, providing a count of the cases characterised by each item. The framework was considered incomplete if one of the items identified in the case study was not covered in a given step. If this is not the case, the framework will be completed and validated against the current literature.

Each of the 24 selected cases underwent a thorough study conducted by multiple readers, each independently completed grid analysis to eliminate any potential influence. Additionally, each reader was required to provide a formal justification for the proposed positioning. The results obtained were compared and analyzed before the final placement of each case in the table.

We encountered no cases that could be characterised using the proposed framework. Additionally, we did not come across any cases that required an item outside the choices of the framework, as indicated by the absence of empty marks in the ‘other’ columns. This provided an initial level of validation based on the literature (see Table ).

6. Discussion

This framework does not necessarily require starting with a specific technology to construct a use case. Instead, it suggests defining a context and a need first, and then identifying the new technology(ies) most relevant to that context through step 5. As such, the framework is not technology-centred. Additionally, defining the agent holds significant importance within the framework because it greatly influences subsequent decisions. By prioritising a human-centred approach, this framework aligns fully with the principles of Industry 5.0.

The majority of the steps in the framework are based on established and recognised works. For instance, Step 2 draws upon the APICS work centre concept (Pittman and Atwater Citation2022), whereas Step 5 utilises the matrix from the research of Rosin et al. (Citation2022). Although the structure of the framework, as outlined in the initial part of the methodology, must be retained, the specific references used to illustrate each step can be replaced by alternatives. It is important to distinguish between the structure and content of a framework. New references can potentially be aligned with the objectives of the users of the framework or offer more recent insights, ensuring an up-to-date framework.

For example, it might be interesting to modify Step 2 with models that consider a broader vision than that of the work centre. Based on a think tank of both academics and industrialists, Bourdu, Péretié, and Richer (Citation2016) analyzed the notion of autonomy at work in the context of several emerging models of work organisation, such as lean management, liberated enterprises, and responsible enterprises. They proposed a model of autonomy at work based on three dimensions that delineate the space of involvement, direct participation, and the ability to influence and decide on the work of an agent: the task, cooperation, or governance.

The framework consists of six distinct steps traditionally approached in a specific order. However, in practice, it is often beneficial to address the generation of use cases through these steps simultaneously. Strong interactions and dependencies exist between steps, and the choices made in one step can potentially affect the decisions made in the previous steps. As a result, the order in which the steps are approached has minimal impact on the outcome as long as there is a continuous feedback loop. This six-step framework can be viewed as multidimensional, with each dimension representing a different aspect of the use-case generation process.

The framework’s validation reveals three configurations worth analyzing:

Configuration 1: an item proposed by the framework is not found in any of the cases studied. This occurred for items such as cybersecurity technology (Step 5), M2M communication (Step 5), and autonomy type 7 (Step 4). This indicates that these specific themes and their implications for autonomy have not been thoroughly explored in existing research. This highlights the potential for further investigation and research in these areas and presents new avenues for future research.

Configuration 2: none of the proposed steps have been identified in the case studied. Steps 3 and 4 of our framework are not detailed by Doyle-Kent and Shanahan (Citation2022). In this case, it was challenging to identify which step in the decision-making process was emphasised because the case study was more concerned with the benefits perceived by people than with studying how employees deal with problems. Similarly, Step 6 of our framework is not detailed in Helm, Malikova, and Kembro (Citation2023), Lo (Citation2023), Nourmohammadi, Fathi, and Ng (Citation2022), Ordieres-Meré, Gutierrez, and Villalba-Díez (Citation2023), and Sit and Lee (Citation2023). In this case, performance is the primary measure, as supported by our framework. Our framework also aims to enhance the analysis by incorporating other variables of interest to provide a more comprehensive understanding.

Configuration 3: one of the items in a given step is identified in all the cases studied. Although this specific configuration was not identified during the validation process, it is worth noting that most articles consistently utilised certain items. For example, the focus on task (Step 2) and enhanced capture-measure steps (Step 4) are prevalent in many cases. Autonomy type 1 (Step 3), enhanced gap recognition (Step 4), and behavioural variables (Step 6) were also utilised in over half of the cases. When connected to the previous configuration, this observation suggests that as we approach the final stages of the decision-making process in complex autonomy types, there is a notable scarcity of studies on these aspects. This trend can be attributed to the industrial focus on the initial steps of the decision-making process, leaving fewer resources dedicated to exploring the later stages.

Moreover, the validation of the framework presented in this article lays the foundation for a review of the current state of research on this subject. The number of cases relevant to the validation objectives increased throughout the study period. Therefore, the methodology can be enriched and applied regularly to an updated set of cases. These are the beginning of a typology of use cases to determine the impact of new technologies on the emergence of new models of autonomy.

Although the framework only proposes the design of use cases, it allows users to check the relevance of use cases already published by the scientific community. Thus, it is a valuable tool for understanding current research into the impact of new technologies on decision-making autonomy. Furthermore, it seems relevant to use this framework to analyze case studies already published by the scientific community to build a more complete overview.

7. Conclusion

This study allowed the development of a comprehensive use case development framework to support research on the emergence of new autonomy models in I5.0 in a structured, realistic, and comprehensive manner. Through its six steps, the framework allows precise definition a specific research question that is relevant and coherent with the overall research question, realistic proposals in line with the findings and impressions of the subject, and a comprehensive use case context. This is centred on an agent (an operator or team with autonomy regarding tasks or collaboration) and a precise section of the business process model. This involves specific autonomy, in which decision-making steps are enhanced by the group(s) of new technologies. The framework can then define a proper data collection protocol by proposing potentially relevant technologies to retrieve a set of variables of interest, as well as an experimental protocol based on a comparison between several augmented configurations and a control configuration. With its numerous choices, the framework can propose a multitude of use cases resulting from all relevant combinations of the proposed criteria and structure future research on the subject. The framework’s retrospective validation also made it possible to perceive contexts completely unexplored by the current research, opening particularly rich research perspectives. As new technologies continue to emerge, the number of potential cases will increase; however, the framework needs to be updated.

The main contribution of this research is to bridge the gap between techno-centric research and real application context. The proposed framework supports researchers in structuring their work by guiding them in developing a set of relevant and coherent use cases that meet the real needs of industrials and researchers. The proposed framework organises experiments in a systemic way and facilitates analysis of results by making them comparable from one use case to another. The proposed framework guarantees a global research strategy on the impact of new technologies on workers’ autonomy in decision-making. Industry 5.0 places human beings at the centre of processes, and the proposed framework does the same. It invites the researchers who use it to question the impact of new technologies on employees. Beyond that, the impact of new technologies on employees is at the core of their research questions. By structuring the work in this way, this framework ensures that research on new technologies is not limited to a quest for efficiency but always considers employee commitment and well-being. In addition, this model aims to define how new technologies can impact workers’ autonomy in decision-making. This model structures it in this sense by inviting the researcher who uses it to question the nature and sharing of decisions between the employee and the technology.

The main theoretical implication of this study is that it is the first industrial use case development framework. This framework can identify the types of use cases or assist in designing and characterising them depending on the objectives and research questions being addressed. From the perspective of managerial implications, this model can serve as a guide for the characterisation of the problem posed by professionals, its resolution, and the verification of effects through a contextualised use case to decide whether to adopt the group(s) of new technologies targeted.

However, the proposed framework has several limitations. The main limitation is that the framework was built on the case study work of Yin (Citation2018), as there is no use case development framework, to our knowledge, in an industrial context. Moreover, few use cases of I5.0 have been produced; however, a trend is emerging. Although the proposed framework enables the structuring of a set of coherent use cases for the same overall research question, the time required to complete each use case remains an obstacle to implementing this type of approach. This is particularly true for Step 1 of the framework: it does not allow the identification of the business process models most likely to serve as ‘standard’ or common support for several use cases. Furthermore, during Step 5, the number of cases that could potentially be studied was extremely large, and it was essential to prioritise them. According to a panel of academic and industrial experts, the Delphi–Régnier study used to generate an example of a relevance matrix made it possible to target technologies with the most significant consensus or dissensus to enhance the decision-making process. Finally, Step 6 of the framework does not specify how to optimise the time required to acquire, process, and analyze the collected data, which can be time-consuming.

These perspectives aim to reinforce the framework by eliminating the identified limitations. The first involves conducting experimental research in the form of use cases to reinforce or enrich the validity of the framework. Several of these studies are currently underway and focus on AI, augmented reality, and a combination of the two. During Step 2, most use cases studied today concern operators who apply it to their tasks. It would now seem advisable to structure a research plan to propose a set of use cases on the other dimension of autonomy or perform a crossover approach on the two dimensions of autonomy. To improve Step 5, a complement to the Delphi–Régnier study could target company managers’ expectations and priorities. This made it possible to specify the study cases that were most expected in the short and medium terms. Particularly, it would be interesting to determine the dimensions (Step 2) and type (Step 3) of autonomy. The proposed framework can be strengthened with a relevance score model that permits the comparison of potential cases, thus prioritising future research experimental conditions. The necessary measures and coefficients to support such an evaluation must be determined based on a thorough and cross-referenced literature study and validated by experts and practitioners. Given the limitations of Step 6, a complementary study could be conducted to identify new technologies that facilitate the acquisition, processing, and analysis of these data. Moreover, designing the use cases requires a multidisciplinary approach. However, this framework does not specify the skills required to design and implement each use case. Furthermore, the framework could become the foundation of a more global meta-analysis tool. This would enable cross-analysis of different research projects concerning a particular group of technologies, for example. This could eventually become a framework for developing use cases and comparing research results. Finally, following the methodology used in this study, it is possible to design other frameworks following the same structure. This would involve modifying steps 3 and 4, currently strongly linked to employee autonomy, to integrate models concerning other industrial concepts, such as proactivity, collaboration, or commitment.

It would be interesting to examine proactivity in the context of autonomy. Parker, Williams, and Turner (Citation2006) define proactivity as a self-initiated anticipatory action aimed at changing and improving one’s situation or oneself. It turns out that worker autonomy ‘contributes to the prediction of proactive behaviour’ (Parker and Collins Citation2010). Consequently, if new technologies have a potential impact on the emergence of new models of autonomy for workers, it seems relevant to ask whether these technologies can extend their impact to proactive behaviour of workers.

In conclusion, it would seem interesting to initiate a literature review on a broader perimeter in terms of discipline and innovations (e.g. Industry 4.0), aimed at taking the state-of-the-art in research and reinforcing the model.

Disclosure statement

No potential conflict of interest was reported by the author(s).

Data availability statement

The data supporting this study’s findings are available from the corresponding author, AG, upon reasonable request.

Additional information

Notes on contributors

Alexandre Goujon

Alexandre Goujon is professor of industrial management and organisation at Arts et Métiers. He teaches industrial management, lean manufacturing, project management, and decision-making. He is currently working on his doctoral thesis in industrial engineering in the LAMIH-CNRS, the Automatic Control, Computer Science, and Mechanical research laboratory in Paris. His research focuses on the impact of Industry 5.0 technologies on human factors such as autonomy, decision-making and involvement.

Frédéric Rosin

Frédéric Rosin is a professor at Arts et Métiers Institute of Technology within the LISPEN laboratory. He holds a Ph.D. in industrial engineering. His research focuses on new production models combining the benefits of operational excellence and Industry 5.0, with a particular focus on enhancing the decision-making process through new technologies. He is the founder of DynEO, a learning – factory offering training in Lean 5.0 which integrates a factory lab for research activities.

Florian Magnani

Florian Magnani is an Associate Professor at Ecole Centrale Méditerranée, France (formerly Centrale Marseille) and conducts research in operations, industrial organisation and human resources at CERGAM (IAE Aix-Marseille). His current research focuses on the adoption of organisational innovations such as Lean, Industry 4.0, Industry 5.0. He pays special attention to the human dimension of these innovations, through the employee behaviour towards the innovations and the impact of these organisational innovations.

Samir Lamouri

Samir Lamouri received his Ph.D. (1989) from the Institut National Polytechnique of Lorraine (INPL), France. He is currently a full professor at Arts et Métiers Institute of Technology (France). He teaches Production Management, Linear Control, and Simulation. His research is conducted as part of the LAMIH-CNRS, the Automatic Control, Computer Science, and Mechanical research laboratory, also in Paris. His areas of interest include distributed and heterarchical control of discrete event systems (manufacturing, transport, logistics, and services). He is a member of the French CNRS research group, MACS, and author of several publications in the manufacturing domain.

Robert Pellerin

Robert Pellerin is a full professor at Polytechnique Montréal in the Department of Mathematics and Industrial Engineering. He has held the Jarislowsky/AtkinsRéalis Research Chair in Project Management since 2010, and is a member of CIRRELT, IVADO, and the Poly-Industries 4.0 laboratory. Prior to joining the Polytechnique, he led various reengineering and implementation projects of integrated management systems and manufacturing execution systems. His industrial and academic career has always focused on the development of tools dedicated to real-time decision-making. His research focuses on these issues, and he has led multiple research projects involving developing and adopting Industry 5.0 practices in various industrial sectors.

Laurent Joblot

Laurent JOBLOT is a researcher at Arts et Métiers Cluny. He holds a Ph.D. in Industrial Engineering (BIM) from this university. After teaching design and project management, he shifted towards implementing training programmes and teaching LEAN through learning labs. He explores the effects, contributions, and prospects of Industry 4.0 technologies (IA, RA, RV, Digital Twin) on humans factors (motivation, involvement, decision-making) in sectors such as manufacturing and construction.

References

  • Abdous, M.-A., X. Delorme, D. Battini, and S. Berger-Douce. 2022. “Multi-objective Collaborative Assembly Line Design Problem with the Optimisation of Ergonomics and Economics.” International Journal of Production Research 61 (22): 7830–7845. https://doi.org/10.1080/00207543.2022.2153185.
  • Attaran, M., and B. G. Celik. 2023. “Digital Twin: Benefits, Use Cases, Challenges, and Opportunities.” Decision Analytics Journal 6: 100165. https://doi.org/10.1016/j.dajour.2023.100165.
  • Autor, D. H., D. A. Mindell, E. Reynolds, and R. M. Solow. 2022. The Work of the Future: Building Better Jobs in an Age of Intelligent Machines. Cambridge, MA: The MIT Press.
  • Ávila-Gutiérrez, M. J., F. Aguayo-González, and J. R. Lama-Ruiz. 2021. “Framework for the Development of Affective and Smart Manufacturing Systems Using Sensorised Surrogate Models.” Sensors 21 (7): Article 7. https://doi.org/10.3390/s21072274.
  • Betella, A., and P. F. Verschure. 2016. “The Affective Slider: A Digital Self-Assessment Scale for the Measurement of Human Emotions.” PLoS One 11 (2): e0148037. https://doi.org/10.1371/journal.pone.0148037.
  • Bourdu, É, M.-M. Péretié, and M. Richer. 2016. La qualité de vie au travail: Un levier de compétitivité refonder les organisations du travail (1–1). Paris (France): Presses des mines-Transvalor La Fabrique de l’industrie.
  • Brey, P. 1999. “Worker Autonomy and the Drama of Digital Networks in Organizations.” Journal of Business Ethics 22 (1): 15–25. https://doi.org/10.1023/A:1006199816737.
  • Bueno, A., M. Godinho Filho, and A. G. Frank. 2020. “Smart Production Planning and Control in the Industry 4.0 Context: A Systematic Literature Review.” Computers & Industrial Engineering 149: 106774. https://doi.org/10.1016/j.cie.2020.106774.
  • Cagliano, R., F. Canterino, A. Longoni, and E. Bartezzaghi. 2019. “The Interplay Between Smart Manufacturing Technologies and Work Organization: The Role of Technological Complexity.” International Journal of Operations & Production Management 39 (6/7/8): 913–934. https://doi.org/10.1108/IJOPM-01-2019-0093.
  • Calzavara, M., D. Battini, D. Bogataj, F. Sgarbossa, and I. Zennaro. 2020. “Ageing Workforce Management in Manufacturing Systems: State of the art and Future Research Agenda.” International Journal of Production Research 58 (3): 729–747. https://doi.org/10.1080/00207543.2019.1600759.
  • Calzavara, M., M. Faccio, and I. Granata. 2023. “Multi-objective Task Allocation for Collaborative Robot Systems with an Industry 5.0 Human-Centered Perspective.” The International Journal of Advanced Manufacturing Technology 128 (1): 297–314. https://doi.org/10.1007/s00170-023-11673-x.
  • Cirillo, V., M. Rinaldini, J. Staccioli, and M. E. Virgillito. 2021. “Technology vs. Workers: The Case of Italy’s Industry 4.0 Factories.” Structural Change and Economic Dynamics 56: 166–183. https://doi.org/10.1016/j.strueco.2020.09.007.
  • Danjou, C., R. Pellerin, and L. Rivest. 2017. Le passage au numérique: Industrie 4.0: des pistes pour aborder l’ère du numérique et de la connectivité, 26. Québec (Canada): Centre francophone d’informatisation des organisations (CEFRIO). https://espace2.etsmtl.ca/id/eprint/14934/1/le-passage-au-num%C3%A9rique.pdf.
  • Deci, E. L., and R. M. Ryan. 1985. “The General Causality Orientations Scale: Self-Determination in Personality.” Journal of Research in Personality 19 (2): 109–134. https://doi.org/10.1016/0092-6566(85)90023-6.
  • De Jonge, J. 1995. Job autonomy, well-being, and health: A study among Dutch health care workers. Maastricht (Netherlands): Maastricht University.
  • De Pace, F., F. Manuri, A. Sanna, and C. Fornaro. 2020. “A Systematic Review of Augmented Reality Interfaces for Collaborative Industrial Robots.” Computers & Industrial Engineering 149: 106806. https://doi.org/10.1016/j.cie.2020.106806.
  • de Paula Ferreira, W., F. Armellini, L. A. de Santa-Eulalia, and V. Thomasset-Laperrière. 2022. “A Framework for Identifying and Analysing Industry 4.0 Scenarios.” Journal of Manufacturing Systems 65: 192–207. https://doi.org/10.1016/j.jmsy.2022.09.002.
  • de Terssac, G. 2012. “Autonomie et travail.” In Dictionnaire du travail, edited by PUF, 47–53. Toulouse (France): PUF. https://univ-tlse2.hal.science/hal-00846542.
  • Doellken, M., J. Zapata, N. Thomas, and S. Matthiesen. 2021. “Implementing Innovative Gaze Analytic Methods in Design for Manufacturing: A Study on eye Movements in Exploiting Design Guidelines.” Procedia CIRP 100: 415–420. https://doi.org/10.1016/j.procir.2021.05.097.
  • Dornelles, J., A. de, N. F. Ayala, and A. G. Frank. 2022. “Smart Working in Industry 4.0: How Digital Technologies Enhance Manufacturing Workers’ Activities.” Computers & Industrial Engineering 163: 107804. https://doi.org/10.1016/j.cie.2021.107804.
  • Doyle-Kent, M., and P. Kopacek. 2021. “Adoption of Collaborative Robotics in Industry 5.0. An Irish Industry Case Study.” IFAC-PapersOnLine 54 (13): 413–418. https://doi.org/10.1016/j.ifacol.2021.10.483.
  • Doyle-Kent, M., and B. W. Shanahan. 2022. “The Development of a Novel Educational Model to Successfully Upskill Technical Workers for Industry 5.0: Ireland a Case Study.” IFAC-PapersOnLine 55 (39): 425–430. https://doi.org/10.1016/j.ifacol.2022.12.072.
  • Eden, D. 2017. “Field Experiments in Organizations.” Annual Review of Organizational Psychology and Organizational Behavior 4: 91–122. https://doi.org/10.1146/annurev-orgpsych-041015-062400.
  • Enang, E., M. Bashiri, and D. Jarvis. 2023. “Exploring the Transition from Techno Centric Industry 4.0 Towards Value Centric Industry 5.0: A Systematic Literature Review.” International Journal of Production Research 61 (22): 7866–7902. https://doi.org/10.1080/00207543.2023.2221344.
  • Eslami, M. H., H. Jafari, L. Achtenhagen, J. Carlbäck, and A. Wong. 2021. “Financial Performance and Supply Chain Dynamic Capabilities: The Moderating Role of Industry 4.0 Technologies.” International Journal of Production Research, 1–18. https://doi.org/10.1080/00207543.2021.1966850.
  • European Commission. Directorate General for Research and Innovation. 2021. Industry 5.0: Towards a Sustainable, Human Centric and Resilient European Industry. Brussels (Belgium): Publications Office.
  • Everaere, C. 2007. “Proposition d’un outil d’évaluation de l’autonomie dans le travail.” Revue française de gestion 34 (11): 45–59. https://doi.org/10.3166/rfg.180.45-59.
  • Fonda, E., and A. Meneghetti. 2022. “The Human-Centric SMED.” Sustainability 14 (1): Article 1. https://doi.org/10.3390/su14010514.
  • Frazzon, E. M., J. Hartmann, T. Makuschewitz, and B. Scholz-Reiter. 2013. “Towards Socio-Cyber-Physical Systems in Production Networks.” Procedia CIRP 7: 49–54. https://doi.org/10.1016/j.procir.2013.05.009.
  • Gagné, M., and E. L. Deci. 2005. “Self-Determination Theory and Work Motivation.” Journal of Organizational Behavior 26: 331–362. https://doi.org/10.1002/job.322.
  • Gao, Y., M. Li, and S. Sun. 2022. “Field Experiments in Operations Management.” Journal of Operations Management, 676–701. https://doi.org/10.1002/joom.1240.
  • Gao, N., W. Shao, M. S. Rahaman, and F. D. Salim. 2020. “N-gage: Predicting in-Class Emotional, Behavioural and Cognitive Engagement in the Wild.” Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies 4 (3): 1–26.
  • Garcia, A., M. Quartulli, I. G. Olaizola, and I. Barandiaran. 2022. Relevance of Visualization and Interaction Technologies for Industry 5.0. 3214. Valencia (Spain): Scopus.
  • Golan, M., Y. Cohen, and G. Singer. 2020. “A Framework for Operator – Workstation Interaction in Industry 4.0.” International Journal of Production Research 58 (8): 2421–2432. https://doi.org/10.1080/00207543.2019.1639842.
  • Grosse, E. H. 2023. “Application of Supportive and Substitutive Technologies in Manual Warehouse Order Picking: A Content Analysis.” International Journal of Production Research, 1–20. https://doi.org/10.1080/00207543.2023.2169383.
  • Guay, F., R. J. Vallerand, and C. Blanchard. 2000. “On the Assessment of Situational Intrinsic and Extrinsic Motivation: The Situational Motivation Scale (SIMS).” Motivation and Emotion 24: 175–213. https://doi.org/10.1023/A:1005614228250.
  • Guo, D., P. Lin, Z. Lyu, S. Ling, M. Li, G. Q. Huang, and Y. Rong. 2019. “Towards Assembly 4.0: Graduation Intelligent Manufacturing System for Fixed-Position Assembly Islands.” IFAC-PapersOnLine 52 (13): 1513–1518. https://doi.org/10.1016/j.ifacol.2019.11.414.
  • Hackman, J. R., and G. R. Oldham. 1976. “Motivation Through the Design of Work: Test of a Theory.” Organizational Behavior and Human Performance 16: 250–279. https://doi.org/10.1016/0030-5073(76)90016-7.
  • Hart, S. G., and L. E. Staveland. 1988. “Development of NASA-TLX (Task Load Index): Results of Empirical and Theoretical Research.” In Advances in Psychology, Vol. 52, edited by P. A. Hancock and N. Meshkati, 139–183. Amsterdam (Netherlands): Elsevier.
  • Helm, M., A. Malikova, and J. Kembro. 2023. “Rooting out the Root Causes of Order Fulfilment Errors: A Multiple Case Study.” International Journal of Production Research, https://doi.org/10.1080/00207543.2023.2251060.
  • Highhouse, S. 2009. “Designing Experiments That Generalize.” Organizational Research Methods 12 (3): 554–566. https://doi.org/10.1177/1094428107300396.
  • Ivanov, D. 2022. “The Industry 5.0 Framework: Viability-Based Integration of the Resilience, Sustainability, and Human-Centricity Perspectives.” International Journal of Production Research 61: 1683–1695. https://doi.org/10.1080/00207543.2022.2118892.
  • Jacobson, I., I. Spence, and K. Bittner. 2011. Use-Case 2.0, The Guide to Succeeding with Use Cases. Ivar Jacobson International.
  • Joblot, L., F. Magnani, F. Rosin, M. Passalacqua, and R. Pellerin. 2023. Protocole expérimental visant l’étude de l’IA centrée sur l’humain dans le contexte de l’Industrie 5.0: Application en réalité augmentée.
  • Kaasinen, E., F. Schmalfuß, C. Özturk, S. Aromaa, M. Boubekeur, J. Heilala, P. Heikkilä, et al. 2020. “Empowering and Engaging Industrial Workers with Operator 4.0 Solutions.” Computers & Industrial Engineering 139: 105678. https://doi.org/10.1016/j.cie.2019.01.052.
  • Katok, E. 2011. “Laboratory Experiments in Operations Management.” In Transforming Research Into Action, edited by Informs, 15–35. Catonsville (Maryland): INFORMS.
  • Kim, H.-G., E.-J. Cheon, D.-S. Bai, Y. H. Lee, and B.-H. Koo. 2018. “Stress and Heart Rate Variability: A Meta-Analysis and Review of the Literature.” Psychiatry Investigation 15 (3): 235. https://doi.org/10.30773/pi.2017.08.17.
  • Klein, G. 2008. “Naturalistic Decision Making.” Human Factors: The Journal of the Human Factors and Ergonomics Society 50 (3): 456–460. https://doi.org/10.1518/001872008X288385.
  • Kumar, R., P. Gupta, S. Singh, and D. Jain. 2021. “Human Empowerment by Industry 5.0 in Digital Era: Analysis of Enablers.” In Advances in Industrial and Production Engineering, edited by R. K. Phanden, K. Mathiyazhagan, R. Kumar, and J. Paulo Davim, 401–410. Singapore (Singapore): Springer.
  • Lai, Z.-H., W. Tao, M. C. Leu, and Z. Yin. 2020. “Smart Augmented Reality Instructional System for Mechanical Assembly Towards Worker-Centered Intelligent Manufacturing.” Journal of Manufacturing Systems 55: 69–81. https://doi.org/10.1016/j.jmsy.2020.02.010.
  • Lesch, V., M. Züfle, A. Bauer, L. Iffländer, C. Krupitzer, and S. Kounev. 2023. “A Literature Review of IoT and CPS – What They are, and What They are not.” Journal of Systems and Software 200: 111631. https://doi.org/10.1016/j.jss.2023.111631.
  • Levenstein, S., C. Prantera, V. Varvo, M. L. Scribano, E. Berto, C. Luzi, and A. Andreoli. 1993. “Development of the Perceived Stress Questionnaire: A new Tool for Psychosomatic Research.” Journal of Psychosomatic Research 37 (1): 19–32. https://doi.org/10.1016/0022-3999(93)90120-5.
  • Lo, H.-W. 2023. “A Data-Driven Decision Support System for Sustainable Supplier Evaluation in the Industry 5.0 Era: A Case Study for Medical Equipment Manufacturing.” Advanced Engineering Informatics 56, https://doi.org/10.1016/j.aei.2023.101998.
  • Longo, F., L. Nicoletti, and A. Padovano. 2017. “Smart Operators in Industry 4.0: A Human-Centered Approach to Enhance Operators’ Capabilities and Competencies Within the new Smart Factory Context.” Computers & Industrial Engineering 113: 144–159. https://doi.org/10.1016/j.cie.2017.09.016.
  • Longo, F., A. Padovano, and S. Umbrello. 2020. “Value-oriented and Ethical Technology Engineering in Industry 5.0: A Human-Centric Perspective for the Design of the Factory of the Future.” Applied Sciences 10 (12): 4182–4125. https://doi.org/10.3390/APP10124182.
  • Macal, C. M., and M. J. North. 2009. “Tutorial on Agent-Based Modelling and Simulation.” Journal of Simulation 4 (3): 151–162. https://doi.org/10.1057/jos.2010.3.
  • Maddikunta, P. K. R., Q. V. Pham, B. Prabadevi, N. Deepa, K. Dev, T. R. Gadekallu, R. Ruby, and M. Liyanage. 2021. “Industry 5.0: A Survey on Enabling Technologies and Potential Applications.” Journal of Industrial Information Integration 26, https://doi.org/10.1016/j.jii.2021.100257.
  • Maghazei, O., M. A. Lewis, and T. H. Netland. 2022. “Emerging Technologies and the use Case: A Multi-Year Study of Drone Adoption.” Journal of Operations Management 68 (6–7): 560–591. https://doi.org/10.1002/joom.1196.
  • Maltseva Reiby, K. 2020. “Wearables in the Workplace: The Brave new World of Employee Engagement.” Business Horizons 63, https://doi.org/10.1016/j.bushor.2020.03.007.
  • Margherita, E. G., and A. M. Braccini. 2021. “Managing Industry 4.0 Automation for Fair Ethical Business Development: A Single Case Study.” Technological Forecasting and Social Change 172: 121048. https://doi.org/10.1016/j.techfore.2021.121048.
  • Martin, C., J. Cegarra, and P. Averty. 2011. “Analysis of Mental Workload During En-Route Air Traffic Control Task Execution Based on Eye-Tracking Technique.” In Engineering Psychology and Cognitive Ergonomics, edited by D. Harris, 592–597. Berlin, Heidelberg (Germany): Springer.
  • Meindl, B., N. F. Ayala, J. Mendonça, and A. G. Frank. 2021. “The Four Smarts of Industry 4.0: Evolution of ten Years of Research and Future Perspectives.” Technological Forecasting and Social Change 168: 120784. https://doi.org/10.1016/j.techfore.2021.120784.
  • Meyer, C. B. 2001. “A Case in Case Study Methodology.” Field Methods 13 (4): 329–352. https://doi.org/10.1177/1525822X0101300402.
  • Mintzberg, H., D. Raisinghani, and A. Théorêt. 1976. “The Structure of “Unstructured” Decision Processes.” Administrative Science Quarterly 21 (2): 246–275. https://doi.org/10.2307/2392045.
  • Mo, F., F. M. Monetti, A. Torayev, H. U. Rehman, J. A. Mulet Alberola, N. Rea Minango, H. N. Nguyen, A. Maffei, and J. C. Chaplin. 2023. “A Maturity Model for the Autonomy of Manufacturing Systems.” The International Journal of Advanced Manufacturing Technology 126 (1): 405–428. https://doi.org/10.1007/s00170-023-10910-7.
  • Moencks, M., E. Roth, T. Bohné, D. Romero, and J. Stahre. 2022. “Augmented Workforce Canvas: A Management Tool for Guiding Human-Centric, Value-Driven Human-Technology Integration in Industry.” Computers & Industrial Engineering 163: 107803. https://doi.org/10.1016/j.cie.2021.107803.
  • Moya, A., L. Bastida, P. Aguirrezabal, M. Pantano, and P. Abril-Jiménez. 2023. “Augmented Reality for Supporting Workers in Human–Robot Collaboration.” Multimodal Technologies and Interaction 7 (4), https://doi.org/10.3390/mti7040040.
  • Müller, J. 2020. Enabling Technologies for Industry 5.0: Results of a Workshop with Europe’s Technology Leaders. Luxembourg (Luxembourg): Publications Office of the European Union.
  • Muñoz, A., X. Mahiques, J. E. Solanes, A. Martí, L. Gracia, and J. Tornero. 2019. “Mixed Reality-Based User Interface for Quality Control Inspection of car Body Surfaces.” Journal of Manufacturing Systems 53: 75–92. https://doi.org/10.1016/j.jmsy.2019.08.004.
  • Nahavandi, S. 2019. “Industry 5.0 – A Human-Centric Solution.” Sustainability 11 (16): Article 16. https://doi.org/10.3390/su11164371.
  • Nguyen Ngoc, H., G. Lasa, and I. Iriarte. 2022. “Human-centred Design in Industry 4.0: Case Study Review and Opportunities for Future Research.” Journal of Intelligent Manufacturing 33 (1): 35–76. https://doi.org/10.1007/s10845-021-01796-x.
  • Nourmohammadi, A., M. Fathi, and A. H. C. Ng. 2022. “Balancing and Scheduling Assembly Lines with Human-Robot Collaboration Tasks.” Computers & Operations Research 140: 105674. https://doi.org/10.1016/j.cor.2021.105674.
  • Object Management Group. 2023. Decision Model and Notation Specification. Milford, MA: Object Management Group. https://www.omg.org/spec/DMN/1.4/PDF.
  • Oldham, G. R., and A. Cummings. 1996. “Employee Creativity: Personal and Contextual Factors at Work.” Academy of Management Journal 39 (3): 607–634. https://doi.org/10.2307/256657.
  • Olhager, J., and A. Feldmann. 2022. “Linking Plant Roles and Operations Strategy Decision-Making Autonomy in International Manufacturing Networks.” International Journal of Production Research 60 (1): 242–255. https://doi.org/10.1080/00207543.2021.1991026.
  • Ordieres-Meré, J., M. Gutierrez, and J. Villalba-Díez. 2023. “Toward the Industry 5.0 Paradigm: Increasing Value Creation Through the Robust Integration of Humans and Machines.” Computers in Industry 150: 103947. https://doi.org/10.1016/j.compind.2023.103947.
  • Orso, V., R. Ziviani, G. Bacchiega, G. Bondani, A. Spagnolli, and L. Gamberini. 2022. “Employee-centric Innovation: Integrating Participatory Design and Video-Analysis to Foster the Transition to Industry 5.0.” Computers & Industrial Engineering 173: 108661. https://doi.org/10.1016/j.cie.2022.108661.
  • Panagou, S., W. P. Neumann, and F. Fruggiero. 2023. “A Scoping Review of Human Robot Interaction Research Towards Industry 5.0 Human-Centric Workplaces.” International Journal of Production Research, 1–17. https://doi.org/10.1080/00207543.2023.2172473.
  • Park, Y. W., and J. Shintaku. 2022. “Sustainable Human–Machine Collaborations in Digital Transformation Technologies Adoption: A Comparative Case Study of Japan and Germany.” Sustainability 14 (17): Article 17. https://doi.org/10.3390/su141710583.
  • Parker, S. K., and C. G. Collins. 2010. “Taking Stock: Integrating and Differentiating Multiple Proactive Behaviors.” Journal of Management 36 (3): 633–662. https://doi.org/10.1177/0149206308321554.
  • Parker, S. K., H. M. Williams, and N. Turner. 2006. “Modeling the Antecedents of Proactive Behavior at Work.” Journal of Applied Psychology 91 (3): 636–652. https://doi.org/10.1037/0021-9010.91.3.636.
  • Passalacqua, M., R. Pellerin, P. Doyon-Poulin, L. Del-Aguila, J. Boasen, and P.-M. Léger. 2022. “Human-Centred AI in the Age of Industry 5.0: A Systematic Review Protocol.” HCI International 2022 – Late Breaking Papers: Interacting with eXtended Reality and Artificial Intelligence: 24th International Conference on Human-Computer Interaction, HCII 2022, Virtual Event, June 26 – July 1, 2022, Proceedings, 483–492.
  • Passalacqua, M., R. Pellerin, F. Magnani, L. Joblot, E. Yahia, F. Rosin, and P.-M. Léger. 2023. L’origine de l’objectif est-elle importante? Effets motivationnels d’objectifs autodéfinis en production.
  • Pereira, A. C., A. C. Alves, and P. Arezes. 2023. “Augmented Reality in a Lean Workplace at Smart Factories: A Case Study.” Applied Sciences (Switzerland) 13 (16), https://doi.org/10.3390/app13169120.
  • Peruzzini, M., F. Grandi, and M. Pellicciari. 2017. “Benchmarking of Tools for User Experience Analysis in Industry 4.0.” Procedia Manufacturing 11: 806–813. https://doi.org/10.1016/j.promfg.2017.07.182.
  • Peruzzini, M., F. Grandi, and M. Pellicciari. 2020. “Exploring the Potential of Operator 4.0 Interface and Monitoring.” Computers & Industrial Engineering 139: 105600. https://doi.org/10.1016/j.cie.2018.12.047.
  • Pinzone, M., F. Albè, D. Orlandelli, I. Barletta, C. Berlin, B. Johansson, and M. Taisch. 2020. “A Framework for Operative and Social Sustainability Functionalities in Human-Centric Cyber-Physical Production Systems.” Computers & Industrial Engineering 139: 105132. https://doi.org/10.1016/j.cie.2018.03.028.
  • Piquet, A. 2009. Guide pratique du travail collaboratif: Théories, méthodes et outils au service de la collaboration. https://www.a-brest.net/IMG/pdf/Guide_pratique_du_travail_collaboratif.pdf.
  • Pittman, P. H., and J. B. Atwater. 2022. ASCM Supply Chain Dictionary, 17th Edition: Vol. Seventeenth edition. https://learn.ascm.org/apex/scormanywhere__SCORM_Player?FAqGV66Nj7RqVQmX7Q%2FQyfyTRwQfpbMavclgeCw9ORdUeifhwoVy%2FLhLIEz9U%2B2w%2BqsyzlVgMphlGtJbKdZZvQ%3D%3D.
  • Rahman, N. A., J. Muda, M. F. Mohammad, F. Ahmad, S. A. Rahim, and F. Mayor-Vitoria. 2019. “Digitalization and Leap Frogging Strategy Among the Supply Chain Member: Facing GIG Economy and Why Should Logistics Players Care?” International Journal of Supply Chain Management 8: 1042–1048. https://www.semanticscholar.org/paper/Digitalization-and-Leap-Frogging-Strategy-Among-the-Rahman-Muda/99b42269db1918186103bb742118604608ec8593.
  • Roblek, V., M. Meško, and A. Krapež. 2016. “A Complex View of Industry 4.0.” SAGE Open 6 (2): 2158244016653987. https://doi.org/10.1177/2158244016653987.
  • Rosin, F., P. Forget, S. Lamouri, and R. Pellerin. 2021. “Impact of Industry 4.0 on Decision-Making in an Operational Context.” Advances in Production Engineering & Management 16 (4): 500–514. https://doi.org/10.14743/apem2021.4.416.
  • Rosin, F., P. Forget, S. Lamouri, and R. Pellerin. 2022. “Enhancing the Decision-Making Process Through Industry 4.0 Technologies.” Sustainability 14 (1): Article 1. https://doi.org/10.3390/su14010461.
  • Rožanec, J. M., I. Novalija, P. Zajec, K. Kenda, H. Tavakoli Ghinani, S. Suh, E. Veliou, et al. 2022. “Human-centric Artificial Intelligence Architecture for Industry 5.0 Applications.” International Journal of Production Research 61 (0): 6847–6872. https://doi.org/10.1080/00207543.2022.2138611.
  • Ruiz Zúñiga, E., C. Barrera Diaz, A. Navarro, A. Ng, T. Hirose, and H. Nomoto. 2022. Reconfiguration Assessment for Production Volume Changes Using Discrete-Event Simulation: A Large-Size Highly-Customized Product Case Study.
  • Rüßmann, M., M. Lorenz, P. Gerbert, M. Waldner, J. Justus, P. Engel, and M. Harnisch. 2015. “Industry 4.0: The Future of Productivity and Growth in Manufacturing Industries.” Boston Consult. Group 9: 54–89.
  • Saihi, A., M. Awad, and M. Ben-Daya. 2021. “Quality 4.0: Leveraging Industry 4.0 technologies to improve quality management practices – a systematic review.” International Journal of Quality & Reliability Management 40 (2): 628–650. https://doi.org/10.1108/IJQRM-09-2021-0305.
  • Schaufeli, W. B., A. B. Bakker, and M. Salanova. 2003. Utrecht Work Engagement Scale-9. Utrecht (Netherlands): Educational and Psychological Measurement.
  • Schramm, W. 1971. Notes on Case Studies of Instructional Media Projects. https://eric.ed.gov/?id=ed092145.
  • Schuh, G., T. Potente, R. Varandani, C. Hausberg, and B. Fränken. 2014. “Collaboration Moves Productivity to the Next Level.” Procedia CIRP 17: 3–8. https://doi.org/10.1016/j.procir.2014.02.037.
  • Simeone, A., R. Grant, W. Ye, and A. Caggiano. 2023. “A Human-Cyber-Physical System for Operator 5.0 Smart Risk Assessment.” The International Journal of Advanced Manufacturing Technology 129 (5–6): 2763–2782. https://doi.org/10.1007/s00170-023-12481-z.
  • Simon, H. A. 1960. The new Science of Management Decision. 1st ed., Vol. 3. New York : Harper & Brothers.
  • Sit, S. K. H., and C. K. M. Lee. 2023. “Design of a Digital Twin in Low-Volume, High-Mix Job Allocation and Scheduling for Achieving Mass Personalization.” Systems 11 (9), https://doi.org/10.3390/systems11090454.
  • Spreitzer, G. M. 1995. “Psychological Empowerment in the Workplace: Dimensions, Measurement, and Validation.” Academy of Management Journal 38 (5): 1442–1465. https://doi.org/10.2307/256865.
  • Sun, S., X. Zheng, B. Gong, J. García Paredes, and J. Ordieres-Meré. 2020. “Healthy Operator 4.0: A Human Cyber–Physical System Architecture for Smart Workplaces.” Sensors 20 (7): Article 7. https://doi.org/10.3390/s20072011.
  • Szajna, A., and M. Kostrzewski. 2022. “AR-AI Tools as a Response to High Employee Turnover and Shortages in Manufacturing During Regular, Pandemic, and War Times.” Sustainability 14 (11): Article 6729. https://doi.org/10.3390/su14116729.
  • Tao, W., Z.-H. Lai, M. C. Leu, Z. Yin, and R. Qin. 2019. “A Self-Aware and Active-Guiding Training & Assistant System for Worker-Centered Intelligent Manufacturing.” Manufacturing Letters 21: 45–49. https://doi.org/10.1016/j.mfglet.2019.08.003.
  • Ulutas, B. H., N. F. ��zkan, and R. Michalski. 2020. “Application of Hidden Markov Models to eye Tracking Data Analysis of Visual Quality Inspection Operations.” Central European Journal of Operations Research 28 (2): 761–777. https://doi.org/10.1007/s10100-019-00628-x.
  • Uva, A. E., M. Gattullo, V. M. Manghisi, D. Spagnulo, G. L. Cascella, and M. Fiorentino. 2018. “Evaluating the Effectiveness of Spatial Augmented Reality in Smart Manufacturing: A Solution for Manual Working Stations.” The International Journal of Advanced Manufacturing Technology 94 (1): 509–521. https://doi.org/10.1007/s00170-017-0846-4.
  • van Oudenhoven, B., P. Van de Calseyde, R. Basten, and E. Demerouti. 2022. “Predictive Maintenance for Industry 5.0: Behavioural Inquiries from a Work System Perspective.” International Journal of Production Research 61 (22): 7846–7865. https://doi.org/10.1080/00207543.2022.2154403.
  • Verrier, G., and N. Bourgeois. 2016. Faut-il libérer l’entreprise?: Confiance, responsabilité et autonomie au travail. Malakoff (France): Dunod.
  • Xu, L. D., E. L. Xu, and L. Li. 2018. “Industry 4.0: State of the Art and Future Trends.” International Journal of Production Research 56 (8): 2941–2962. https://doi.org/10.1080/00207543.2018.1444806.
  • Yin, R. K. 1981a. “The Case Study as a Serious Research Strategy.” Knowledge 3 (1): 97–114. https://doi.org/10.1177/107554708100300106.
  • Yin, R. K. 1981b. “The Case Study Crisis: Some Answers.” Administrative Science Quarterly 26 (1): 58–65. https://doi.org/10.2307/2392599.
  • Yin, R. K. 2018. Case Study Research and Applications: Design and Methods. Sixth edition. Thousand Oaks, CA: SAGE.
  • Zeisel, S. 2020. “A Framework and Use Cases for Big Data in Procurement.” International Journal of Procurement Management 13 (6): 775–793. https://doi.org/10.1504/IJPM.2020.111346.
  • Zheng, T., C. H. Glock, and E. H. Grosse. 2022. “Opportunities for Using Eye Tracking Technology in Manufacturing and Logistics: Systematic Literature Review and Research Agenda.” Computers & Industrial Engineering 171: 108444. https://doi.org/10.1016/j.cie.2022.108444.

Reprints and Corporate Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

To request a reprint or corporate permissions for this article, please click on the relevant link below:

Academic Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

Obtain permissions instantly via Rightslink by clicking on the button below:

If you are unable to obtain permissions via Rightslink, please complete and submit this Permissions form. For more information, please visit our Permissions help page.