1,653
Views
2
CrossRef citations to date
0
Altmetric
Research Article

Are “wrong” models useful? A qualitative study of discrete event simulation modeller stories

, ORCID Icon & ORCID Icon
Pages 594-606 | Received 06 Oct 2021, Accepted 20 Jul 2022, Published online: 18 Aug 2022

ABSTRACT

Little is known about models deemed ”wrong” by their modellers or clients in Operational Research (OR). This paper aims to improve our understanding of “wrong” Discrete Event Simulation (DES) models based on empirical evidence. We interview 22 modellers who describe projects where modelling did not go as expected and explain how they dealt with those situations. This resulted in 54 stories reporting that a model was identified ”wrong” either by the modeller, the client or both. We perform a qualitative text analysis of the stories to identify the factors that define a ”wrong” model as well as potential uses of ”wrong” models. The results show that some models even though considered ”wrong” may still be useful in practice and provide valuable insights to users and modellers. This study offers practical suggestions for users and modellers to consider when dealing with a model that is considered ”wrong”.

1. Introduction

Simulation is a well-established tool used in Operational Research (Tako et al., Citation2010; Kotiadis & Mingers, Citation2014; Robinson, Citation2014; Tako, Citation2015). Yet, the development of a model may be subject to constraints such as time or resource limitations. As a result, this could lead to a model considered “wrong“ by the users (clients/third parties), the modellers or both based on their perspective and understanding. This suggests that the concept of wrongness may be subjective. From our exploration of the literature, we note that there is limited exploration of “wrong“ models, even though there is a general belief that they can be useful (Hodges, Citation1991). Two main questions arise: what constitutes a “wrong” model and whether a ”wrong” model can still be useful. In this study we explore both questions. We note that in this study we consider ”wrong” models as any models that could subjectively be perceived wrong by anyone involved during their development and/or use.Footnote1

To date there is no clear definition of what constitutes a “wrong“ model in the OR field. Some authors refer to the reasons (called here factors) that may lead to “wrong” models, while others refer to specific model characteristics related to wrongness. For example, Robinson (Citation1999) describes experimentation as a source of model inaccuracy in simulation studies. Vennix (Citation1999) and Salt (Citation2008) refer to clients’ biases and beliefs affecting model creation. Gass (Citation1983) and Kleijnen (Citation1995) on the other hand, explain how to build valid models. We thus find a range of different opinions and viewpoints as to what could be a ”wrong” model.

Following, modellers or users could be faced with the following question: if a model is “wrong“, should it be discarded or can it still be of use? There is little evidence in the literature on possible uses of such models as papers usually showcase successful models (Citation2006) with few references to failed attempts. The limited mentions found are mostly theoretical in nature. For example, Hodges (Citation1991) and Hodges and Dewar (Citation1992) explain uses for promotional or training purposes. An experimental study by A. Tako et al. (Citation2020) approaches the topic of models that may be considered “too simple“ for use and provides evidence that simpler versions of a model could still be useful for users. This is a relevant study to this paper as sometimes models may be simplified to the extent that they are considered “wrong”. However, this is the only study that offers some empirical evidence on this topic, where a simple model could be considered as one type of a ”wrong” model. As such, we still need to understand the potential usefulness of ”wrong” models when encountered in practice.

The research presented here aims to shed further light to our understanding of “wrong“ simulation models in practice by categorising reasons of wrongness and potential uses of “wrong“ models based on empirical evidence. This topic has received little attention in the simulation field. For practical reasons we have chosen to focus our empirical work on Discrete Event Simulation (DES) which is the authors’ main field of research. This also ensures that all examined practical cases use the same type of Simulation. We note, however, that the same type of work could be applied for any other simulation approach, such as System Dynamics or Agent-Based Simulation. We interview DES modellers who discuss real-life examples of developing “wrong“ models and how these are treated by the modellers and users. Two objectives are set: to investigate factors of wrongness related to ”wrong” model characteristics, and to explore possible uses. We perform qualitative text analysis which shows evidence that some ”wrong” models may indeed be used in practice and even provide valuable insights. Modellers and practitioners could utilise the findings from our empirical analysis presented in this paper to identify cases where the use of ”wrong” models could be justified and hence ensure that modelling efforts are not completely lost.

The rest of the paper is structured as follows. Section 2 explores OR and Simulation literature referring to definitions and uses of “wrong” models in relation to the two study objectives. Next, the methodology is briefly explained and results of our analysis are presented. Section 5 discusses our findings against existing literature and puts forward potential reasons that such models could be useful. We conclude by considering limitations and future work.

2. Background on “wrong” models

2.1. Definition and possible uses of “wrong” models

A famous quote by George Box states that “all models are wrong, but some are useful“ (Box & Draper, Citation1987). This suggests that even bad models may offer some value to users. We provide an analysis of the concept of ”wrong” models.

“Wrong“ models are rarely presented (Rouwette et al., Citation2002), not allowing learning outcomes to derive from them (Citation2006). Even when mentioned, different authors prefer different definitions. In OR, Hodges (Citation1991) refers to “bad“ and “inadequate“ models. He distinguishes between models that can be validated but haven’t yet (“unvalidated“) and models that simply cannot be validated (“invalidated“). A similar categorisation is found between validatable models used for predictions, and “unvalidatable“ models for other non-predictive uses (Hodges & Dewar, Citation1992). A different clustering is made between “consolidative“ and “exploratory“ models where “consolidative“ is a correctly constructed and validated model while “exploratory“ is a model used for solutions with unavailable data (Citation1993; Hooker, Citation2007). Authors in Simulation papers use terms like ”false” or ”incorrect” (Bankes, Citation1998; Citation2007) or simply ”wrong” (Hinkkanen et al., Citation1995) to denote a problematic model. Bankes (Citation1998) offers further descriptions like “demonstrably incorrect”, “strongly predictive”, and “plausible” or “weakly predictive”. In another work (Phillips, Citation1984), it is hinted that a model is either “at fault” (“not sufficient to solve the problem”) or “requisite” (has shown new intuitions).

The use of multiple terms to describe “wrong” models suggests that there is no consolidated definition of “wrong” models nor a consistent approach to how they are dealt with in OR. The existing literature often uses subjective definitions of “wrong” models. In this study, we consider a model “wrong” if any problems can be identified with it. We note that our concept of “wrong” models encapsulates any term found in literature that refers to problems found with the models mentioned.

Nevertheless, it is suggested that even a “wrong” model may be useful. There is limited discussion on this with two main papers presenting possible uses (Hodges, Citation1991; Hodges & Dewar, Citation1992). Their suggested uses can be summarised as follows:

  • promotion or selling,

  • training users when accuracy is not required,

  • storing information within a management system,

  • creating new knowledge on results that don’t have to be exact, and

  • implementing “wrong” models’ knowledge to test hypotheses when precision is not required.Footnote2

Other authors refer to model usefulness even if they do not specify whether/how “wrong“ models may offer benefits. For instance, Bankes (Citation1998) states that “exploratory” and ”weakly predictive” models may assist in decision-making, Mens and Van Gorp (Citation2006) refer to usefulness as a quality that helps understand systems and make decisions. This is also supported by Jessop (Citation2002). Citation1999 comments that usefulness is found in how a model addresses expected problems.

Usefulness of “wrong” models is rarely considered. A study that provides relevant evidence on this topic shows that simpler models may be as useful as their more complex equivalents for decision-making (A. Tako et al., Citation2020), where simpler models can be considered a type of a “wrong” model. Yet, to the best of our knowledge, there has not been any empirical explorations about the use of “wrong” models in OR.

2.2. Characteristics of “wrong” models

Models in OR need to be rational to help with decision-making and learning (Ormerod, Citation2018). Since users learn by connecting model results with their understanding, this creates subjectivity for modellers’ and users’ approach to models (Pidd, Citation2009) and to outcomes’ interpretation (Williams, 1990).

This means that validity of a model – “the process of ensuring that the model is sufficiently accurate for the purpose at hand” (Robinson, Citation2014) – may be different to its credibility which concerns a model’s credential value from clients (Gass, Citation1983). This is further supported in recent literature through the exploration of the idea of trust (Harper et al., Citation2021; Yilmaz & Liu, Citation2020). Yilmaz and Liu (Citation2020) posit that model credibility is not only based on numerical accuracy but also on the extent that model users place trust in models and find them believable. Hence, model credibility depends on the cognitive aims and interests of its intended users (Gelfert, Citation2019). Harper et al. (Citation2021) explore trust as an integral part of all model development stages between the stakeholders, modeller and model, highlighting a fundamental need for the clients to acquire confidence from the insights gained from this process. Robinson (Citation2002a) makes a similar observation about the importance of process, as well as content, in forming opinions about the credibility of simulation study.

Following the above, the interpretation of a model being considered “wrong“ can be subjective. We use the framework developed by Robinson (Citation2014) to help us identify the characteristics of a “wrong“ model based on two main aspects: the process followed in developing the model and the content of a model. The characteristics are defined based on the authors’ interpretation of the references found in the existing literature. Hence, we identify two main categories of factors that lead to “wrong” models: process-related factors which refer to the underlying reasons a model may be ”wrongly” developed during a modelling study (model development process), and content-related factors which are related to elements of the model that could be considered ”wrong”. In this research, factors are split under distinctive common model traits called model characteristics. These characteristics were created and assigned to either category based on our analysis and understanding of the existing literature which relates to modelling process and implementation. We provide at the footnote of a brief operational definition for each of these characteristics.

Table 1. Categorisation of factors of wrongness from the literature analysis.

2.2.1. Process-related characteristics

The first category regards process-related factors under four model characteristics: validation, model development, decision-making related to model stakeholders, and technical aspects. Each one is next explained in more detail.

Validation is a very important part of modelling to ensure that a model corresponds to its purpose (Robinson, Citation2014). If validation is not proper, the model may be considered “wrong“. Within OR, Hodges (Citation1991) and Hodges and Dewar (Citation1992) separate models based on validation as mentioned above – if validation is infeasible or difficult to occur, the model may be considered “wrong“. In Simulation, improper validation is another concern: if validation is done “wrongly“ it could lead to a ”wrong” model (Sargent, Citation1991). This is related to experimentation as a source for model inaccuracy as it may produce deductions based on insufficient model replications and result analysis (Robinson, Citation1999). Kleijnen (Citation1995) proposes that lack of validation may offer wrong decisions. This can be supported by the ”black-box mistake” idea where a model’s outcomes are not questioned because they are considered inaccessible (Salt, Citation2008) leading to a ”wrong” model being developed. Still, proper validation is not the only adequate premise to avoid model wrongness (Kleijnen, Citation1995) as we suggested earlier when mentioning credibility.

Model development is a set of reasons related to factors of limitation or restriction like time or funding (Balci et al., Citation2002). Indeed, if resources are limited or run out during a project, a model may be developed hastily or not even fully. Short-term focus is also included (Chussil, Citation2005) as a model could become obsolete after some time if during development we do not consider its use for the future. Certain more theoretical factors like unmatched expectations of technology that OR and Simulation didn’t meet (Bankes, Citation1998; Kolkman et al., Citation2016; Citation2015) are also encountered.

Decision-making related to model stakeholders is commonly connected to model wrongness. It regards decisions of those involved in model development and can be attributed to clients, modellers or interactions (dynamics) between them. In view of clients, a person may call a model “wrong“ due to empirical criteria (wishful thinking, bias, overconfidence) or bad assumptions (Chussil, Citation2005). In Simulation, we find analysis by clients (Citation2007), unquestionable belief in a model (Salt, Citation2008), wrong model selection instead of alternative models that may have been better (Citation2007) or wrong comparison with the system (Brooks & Tobias, Citation1996). Additionally, training and experience of users (Kolkman et al., Citation2016), methodolatry, lack of willingness to change a model, and belief that reality can be simulated perfectly (Salt, Citation2008) are cited. Specifically in System Dynamics we find distorted memory, professional background and position (Vennix, Citation1999). Equally, a model may also be developed ”wrongly” due to modellers’ actions. Factors include lack of objectivity from modellers (Citation1993), modelling only the easy parts of a system (Bankes, Citation1998), and even possible lack of skill (Robinson & Pidd, Citation1998). Lastly, group dynamics can also be connected to wrongness: different interests of clients and modellers (Hodges, Citation1991), lack of customer inclusion or absence of competition (Chussil, Citation2005), and team member communication (Vennix, Citation1999).

Technical aspects is the last characteristic here. Having data problems may lead to a “wrong” model as its development basis would be inaccurate. This is referenced both in OR as contradiction of existing data leading to bad outcomes (Hodges, Citation1991) and in Simulation as problems arising from sampling data (Citation2007). Also, components’ performance and technology appear in literature. In OR, a model may have good partial components but considered bad in its totality (Hodges, Citation1991) and in Simulation a model may be lacking a component or a variable (Pidd, Citation2004) or its components’ performance may fail (Brooks & Tobias, Citation1996).

2.2.2. Content-related characteristics

A similar analysis is performed for factors related to the content of the model.

Level of model detail regards factors in view of a model’s simplification. Simple models may be considered “wrong“ as they may not represent the initial problem adequately (Brooks & Tobias, Citation1996) due to “far abstraction“ (Citation2015). This is the first factor here, addressing that a model may be considered “wrong“ because it is viewed as too simple. In OR, we find that abstraction is needed for a model but it may create inaccuracies (Prieto et al., Citation2012) while in Simulation even a simple model could offer solutions but it could also imply problems (Robinson Citation2015). On the other hand, a model may be considered ”wrong” because it is viewed as too complex as it may include unnecessary details (Citation2017). In OR, Citation1993 suggests that large-size models could be a cause for concern and Goldberg et al. (Citation1990) point that a high level of detail may not actually be required. In Simulation, creation of a more detailed model may be limited by resources (Brooks & Tobias, Citation1996) with relevant factors of wrongness here referencing beliefs that more content leads to better outcomes and connecting good models together also returns a good model (”trifle-worshipping” and ”connectionism”; Salt, Citation2008).

Model results is another model characteristic that is found in cases where wrongness is discussed. If a model is providing wrong outputs, it may be regarded as “wrong“. For instance in OR we find that comparisons with specific values can denote a “wrong” model (Goldberg et al., Citation1990; Hodges & Dewar, Citation1992), while in Simulation inaccurate results and values may also be used to describe a model as ”wrong” (Brooks & Tobias, Citation1996).

Structural or application-related model aspects categorises factors related to issues with the structure and application of a “wrong“ model together. This is because issues with a model’s structure can affect whether the model is applied. In view of structure, Chussil (Citation2005) proposes that it can indeed result in an OR model being considered “wrong“ due to the need for model calculations to be based on well-defined infrastructures. In simulation, Robinson (Citation1999) refers to issues that may be invoked because of bad conceptualisation, while Bankes (Citation1998) supports that if an internal model structure is different to that of a known system, it could lead to the model being considered “wrong”. For model application, reasons identified are less explicit. We find reports that the OR problem may have been translated into the ”wrong” model (Lombaers, Citation1979) and in DES that a wrong problem may have been solved (Kotiadis et al., Citation2014; Law & McComas, Citation1991). This suggests the need for an appropriate interpretation of a problem to avoid subsequently building a ”wrong” model.

summarises the factors of wrongness as categorised under model characteristics.

2.3. The need for exploring “wrong” models

Our analysis of existing literature shows that there is no single definition on the concept of wrongness and that many reasons can contribute to a model being considered “wrong”.

It is our belief that wrongness is the result of subjective decision-making based on how modellers and users perceive the modelling process and model content. This creates the question of what is considered a “wrong“ model. We also note that authors rarely mention a “wrong“ model and that there is a limited discussion on the reasons that lead to a “wrong” model. The use of ”wrong” models is discussed at a theoretical level without practical guidance. There is hence a need for practical guidance in assessing models and whether a ”wrong” model can be useful and how.

This research is a first attempt to address this gap, by categorising factors of wrongness and taking an empirical view in identifying such factors in practice. Furthermore, we aim to provide evidence about the extent to which “wrong” DES models are used in practice and their potential uses.

3. Methodology

3.1. Objectives

The main aim of this paper is to define in practice what is considered a “wrong” model and whether such a model can still be useful. Based on the identified gap (Section 2.3), we set out two main objectives as follows:

  • Objective 1 (O1): to provide a definition of “wrong“ models. O1 addresses the question ”what is considered a ‘wrong’ model”. The research is expected to denote specific factors of wrongness categorised under model characteristics. These factors will be mapped against those found in the literature to establish whether they match modellers’ experience in practice. Factors are expected to be corroborated and additions to be made.

  • Objective 2 (O2): to explore uses of “wrong“ models. O2 addresses whether a ‘wrong’ model can still be useful and identify potential uses“. The first part is establishing whether a “wrong” model may actually be used in practice. If such a model can be considered useful, its uses are further explored. It is expected to find evidence on possible use as well as actual uses of ”wrong” models to accept or refute usefulness for ”wrong” models.

3.2. Choice of research method and data analysis strategy

To select the best research method and way to analyse the collected data, certain requirements are considered.

The two objectives are defined by the need for empirical data to show what consists a “wrong“ model and whether it can be of use. To gather enough information, a method of a qualitative nature would be required, as it would provide detailed information but also allow us to compare outcomes with previous literature suggestions. Interviews are deemed as the most suitable method to gather sufficient data to address the objectives of the study, considering also time restrictions for this research. We conduct semi-structured interviews and ask participants to describe stories reporting examples where DES models deemed ”wrong” were developed, based on their modelling experience. Other researchers have used interviews to research DES modelling practices. Tako (Citation2015) interviewed expert modellers by asking them to speak aloud when they develop DES models. Gogi (Citation2016) interviewed DES modellers to convey stories to explore insight generation from using simulation models in Operational Research. Here we explore factors of wrongness in DES models as well as whether and how ”wrong” models may have been used. We solely focus on DES models, as already explained in Section 1, primarily for practical reasons and due to the lack of previous studies on this topic.

All collected data was transcribed and anonymised. Thematic analysis was undertaken looking for repetitive themes with relevant information to the objectives of this research. (Bryman, Citation2012) content analysis. Interpretation of meaning from text data (Bazeley, Citation2013) was undertaken through rigorous and repetitive exploration of the interview transcripts. The content was grouped into relevant themes and then compared with topics found in the literature. Further details of the qualitative research methodology and data analysis strategy are provided in Appendix A.

3.3. Study participants and interview process

We carry out interviews with experienced DES modellers to explore real-life cases of unsuccessful DES modelling projects. We contacted a number of UK-based simulation companies who helped us to identify potential interviewees. The topic was met with interest, resulting to a total of 22 participants from 6 different companies from the private sector. A total of 54 stories/models were discussed that addressed models of various content, all of which were found relevant to the objectives around “wrong“ models as they contained either issues around modelling or the view that a model was considered ”wrong”. provides a summary of the interviewees in each company, their experience and number of stories shared.

Table 2. Information of the interviewed participants.

Further details about the profile of the participants’ experience, clients’ sectors and problem types they were facing are provided in Appendix B.

Regarding the process, an ethic clearance was performed (Brinkmann & Kvale, Citation2015) to inform every interviewee around confidentiality. Each interview was then held with one of this paper’s researchers and the DES modeller, lasting between 40 and 60 minutes. After exploring each interviewee’s background, they were asked to provide details of projects that did not go well. Questions were used to obtain information for each project on model size, industry of application and project aims. Questions on wrongness followed on the problems with each model and who identified them. These would offer information for O1 on factors related to wrongness in practice. Further questions were asked to clarify model acceptance and use by clients, how that happened, etc. This set of questions aimed to understand usefulness of “wrong“ models, addressing O2 on using “wrong” models. It is noted that due to the nature of semi-structured interviews, specification of questions and additional questions were asked if required to conform by the peculiarities of each situation. The generated data here is primary data collected for this research on the use of ”wrong models” while the timeline for the interviews was approximately 3 months, taking place between November 2018 and January 2019.

4. Results from the interviews

4.1. Defining a “wrong” model

The analysis uses the two categories of factors defining “wrong” models as presented in Section 2. It is noted that although all stories came from the modellers’ descriptions, some of them referred to clients perceptions of the modelling problems (as narrated by the modellers). This distinction in perception is displayed in . All factors are identified based on the researchers’ interpretation of the interviewees’ narrations.

Table 3. Results on factors of wrongness from the interviews per perspective.

4.1.1. Factors of wrongness under process-related characteristics

The same four model characteristics that contain factors of wrongness for process were encountered at various stories.

The most common set of problems reported in 30 of those stories regards Decision-making related to participants. More specifically, factors attributed to the clients that led to a model being considered “wrong“ concern guesses the clients may have made during development or even lack of planning for creating a model, contradicting decisions, and lack of training or understanding of DES. Other problems include evasion of liability or lack of participation when developing a model, dismissal of certain results due to bias and persistence on unhelpful details. Additionally, disbelief towards a model or mismatch of questions with the agreed scope are considered as possible issues towards building a model. Bad communication or details not communicated, lack of strict specifications, and loss of project control are also referenced. There are also certain factors that are attributed to how the modellers have approached a model that was considered ”wrong” and include lack of experience or expertise, bad presentation of model or outcomes, bad communication or misinterpretation of data, guesses or assumptions alongside any uncertainty on clients’ operations, decision to use too many or unnecessary details, and unrealistic or unsuitable promises made to the client. Group dynamics are also discussed, with the main factors being conflicts of interest, internal or external politics causing discrepancies, insistence for simulation use, lack of collaboration or communication, inclusion or interest for a project, departure of important project members, and different processes followed.

The second most recurring set of factors regards data and technical topics (Technical aspects) connected with 20 stories. In view of data concerns, models are developed “wrong“ due to lack of data or because of the involvement of contradicting, messy, ambiguous or bad data. Other stories reference extreme availability of data causing problems, lack of trust in data, and different data specifications within the same company resulting to discrepancies in analysis. Accordingly, modellers mention issues with components or coding that could have been responsible for ”wrong” models. These regard use of wrong distributions, calculations, components or variables, incomplete models with flaws on mapping/layout and missing or wrong logic that were still put to use, wrong implementation of variables or numbers, model speed concerns, and bad rules or priorities causing performance failure.

The next most referenced set of factors is found under Development aspects. There are nine unique references of such factors that could be connected to developing a model “wrong”. Concerns could be summarised under two main factors of wrongness: time limitations to complete a model and the possible frustration this causes, and, short-term focus of model design including incorrect or continuously changing scope as the reasons for developing a bad model.

Lastly, Validation problems are mentioned in eight stories. The factors regard infeasible, impossible or very difficult validation, improper/bad or failed validation, and lack of validation. Interestingly enough, validation factors are always mentioned alongside other factors of wrongness as the cause for developing a “wrong” model and all of them come from the modellers’ side – meaning that their clients did not complain about this topic.

4.1.2. Factors of wrongness under content-related characteristics

All three model characteristics for content were mentioned at various stories, also encountering factors that could be put under a fourth characteristic.

Starting again with the most commonly referenced factors, we find that Model results are considered crucial for finding a model good or bad. The actual factors mentioned here in 27 stories concern values, results or visuals different from the clients’ reality or expectations, results that are doubted wrongly by the clients, and lack of outputs from the model.

The second set of factors with the most references regards the Structural or application-related model aspects characteristic, mentioned in 17 stories. In view of structure, factors include lack of representativeness for model structure, incomplete models, and models wrongly altered by the client. Application issues concern when the wrong question is addressed, not delivering what is promised due to scope change, and when clients’ demands or internal disagreements are involved.

Level of model detail considers the complexity represented in the model. It is noted that due to confidentiality purposes, we were not able to access the models accounted for in the stories. Instead, the analysis is based on the modeller’s interpretation of a too simple or too complex model, if their level of model detail was reported to be a concern. A total of 13 unique references are encountered: models could be “wrong“ due to being either too simple or too complex, while another factor of wrongness suggests that simplification may not be correct. The finding that both too simple and too complex models may be considered ”wrong” is very interesting especially when considering that opinions could be different even for the same model. Indeed in two cases of our study, a relatively complex model for the modeller is considered as too simple by the clients for their needs.

The analysis of the interviews also brought to light another set of factors that literature did not highlight under the characteristic of Usability. There are only two references in stories, but these are very elaborative to consider it separately. Factors here refer to difficulty to use a model, lack of interest to use a model, different format of results than required, and DES not preferred as a method by clients.

4.1.3. Summary for O1

We found a total of eight different model characteristics populated with factors of wrongness. In various cases, more than one factor was referenced, but correlations exceed our current analysis. The distinction of whether factors are coming from modellers and clients is presented in .

Some interesting points arise: Validation does not seem of high concern for clients, as opposed to Usability that modellers do not reference, while Results are equally split – meaning it is a very important characteristic to account for both sides.

4.2. Exploring uses of “wrong” models

The analysis of the stories revealed that “wrong“ models could actually be used, offering benefits to their users. Additional results on why a ”wrong” model was decided to be used and how it was implemented are presented in Appendix C.

4.2.1. Using a “wrong” model in practice

We first explore the extent to which “wrong” models are used in practice. The following diagram () shows the process followed.

Figure 1. Analysis followed for the 54 stories to consider uses.

Figure 1. Analysis followed for the 54 stories to consider uses.

All 54 stories contained a “wrong“ model. As such, we analysed iteratively what modellers mentioned that happened in each case. We use the following operational definitions for our categorisation: If a model was clearly put to some use or meant to be put to some use by the clients regardless of who considered it ”wrong”, it was noted as used. If there was a clear reference that clients would not use the model, it was considered as not used. Stories without information on the topic were categorised as unclear as their use could not be communicated either because the modeller did not follow up to find out, or, because the client did not offer specifications due to confidentiality. summarises the outcomes.

Table 4. Results on interviews regarding whether models were used.

We notice that almost half of the stories regarded a “wrong” model that was considered used.

4.2.2. Uses of “wrong” models in practice

We now focus on the actual uses of those models to provide an understanding of their role in practice. Six types of uses are identified by the analysis, presented in .

Table 5. Results of interviews regarding uses of models.

The majority of stories (17/25) shows that “wrong“ models were used for decision-making (case descriptions in ). The second type of use found in eight stories regards use of “wrong“ models for hypothesis or scenario testing. Some overlapping with the first category is noted. The third encountered type in seven stories regards use of “wrong“ models to help surface issues. The fourth encountered type in three stories regards use of ”wrong” models for better understanding of the real world or to develop clients’ thinking of a problem. This use is always accompanied by other uses. The fifth encountered type in two stories refers to use of ”wrong” models for promoting or communicating a selling idea. Lastly, the sixth encountered type in one story addresses use of ”wrong” models for storing information or monitoring/controlling processes. summarises types and cases.

4.2.3. Summary for O2

We found that in almost half of the interviews (25/54) a “wrong“ model was utilised or was expected to have been utilised. Their actual uses were categorised over 6 overlapping types based on the literature analysis. These are all actual uses of ”wrong” models in practice. Appendix C holds some additional findings on why those models were used and how they were implemented.

5. Discussion of findings

This novel study investigated the unexplored topic of “wrong” models. We next discuss research contributions and the findings of our interviews against existing literature, highlighting additional insights gained and considerations on how different domains of application can benefit from our analysis.

5.1. Contribution of research

This research has for the first time considered why and how “wrong“ models are used, exploring existing literature and DES practice to find factors that lead to the development of “wrong“ models and identify their potential uses. The main contributions regard categorisation and analysis on wrongness and usefulness both for theory and practice. From a theoretical perspective, this study offers a list of factors leading to “wrong” models and their uses. From a practical perspective the described cases offer empirical evidence to support clients and modellers when faced with a ”wrong” model. The very process of developing a model can offer insights even if the outcome may not be as expected. Furthermore, the list of factors can be considered in practice as a topic of discussion when exploring what may go or have gone wrong during development while the uses can be considered as alternative options for finding benefits even from ”wrong” models.

5.2. Factors of wrongness

To address factors of wrongness in practice, the first objective explored them through real-life cases.

We first explored factors leading to the development of a “wrong” model. We identified characteristics for Validation, Model Development, Decision-making related to model stakeholders, and Technical aspects, matching interview findings to literature. More specifically, when discussing Validation, infeasible, impossible or difficult to be carried out validation align with Hodges (Citation1991) and Hodges and Dewar (Citation1992). Most Model Development factors are encountered in practice, meaning that planning and proper scope are elementary in practice, while Decision-making is again split into issues related to clients, to modellers, and to group dynamics. On the contrary, some factors were not repeated from . For example, Model Development funding restrictions and technological constraints (Bankes, Citation1998; Citation2015) are not explicitly mentioned in stories. For Decision-making factors attributed to modellers, we notice a higher variety in interviews than literature. Also, there were some significant additions like Technical aspects factors on speed of models and problems with coding. We noted that Validation concerns are reported only from modellers which could suggest that clients may not be fully acquainted with the concept to reference it as an issue.

We then explored characteristics related to model content. These are the Level of model detail, Model results, Structural or application-related model aspects, and Usability. It is again noted that most findings were matched to literature. For example, for the Level of model detail, the interviews supported that both too simple and too complex models could be considered “wrong”, while for Model results, numerical citations within literature (e.g Goldberg et al., Citation1990) were corroborated. Still, there were factors from literature not fully matched as for instance conceptual concerns under Structural or application-related aspects probably due to firm scoping techniques that companies adapt. On the contrary, additions did not only regard factors but also a new category, Usability, which was suggested as a characteristic related to how easy a model and its components are considered to use.

To summarise, the study discusses a variety of characteristics of “wrong“ models. Their categorisation was possible under the same two categories of Section 2, confirming and expanding most factors, and adding another characteristic. The most commonly cited factors for modellers and clients were found under the same categories (Decision-making and Model results). Still, the findings show that there can be a difference of opinions between modellers and clients on whether and why a model is ”wrong” with complex dynamics on decision-making. Especially in view of level of model detail, simple and complex models may even reflect opposite opinions.

5.3. Usefulness and potential uses of “wrong” models

Following, our study considers uses of “wrong” models from a practical perspective.

In total, 25 of the 54 examined models were found to have been used in at least some way. This derived from different interviewees and companies – which shows that a variety of models may be considered as inadequate but still useful. This also supports the assumption that using a “wrong” model does not imply that a modeller would willingly offer a bad model to their clients – instead clients may even insist on using a model because they find it credible for their needs or because a modeller showed possible benefits attributed to this model’s use (Appendix C).

Six types of uses were found in practice. The most common use (17/25 stories) regarded Decisions or actions that according to modellers offered benefits to clients. This use resembles the exploitation of “wrong“ models to create knowledge on results that don’t have to be exact (Hodges, Citation1991; Hodges & Dewar, Citation1992). In other words, ”wrong” models seem possible to be considered for decisions if precision is not the main concern.

Next, use for Hypothesis or scenario testing (8/25 stories) was found which is also identified in literature on how “wrong“ models can implement acquired knowledge for hypothesis generation (Hodges, Citation1991; Hodges & Dewar, Citation1992). This shows that ”wrong” models may be useful in testing system interactions even if their numbers are inaccurate.

The next most encountered use (7/25 stories) is deemed to Surface problems within an organisation. This use describes that models may offer insights on operational or decision discrepancies for a company which could be connected to using Soft OR for initiating discussions (Mingers, Citation2003) and using simulation to create conversations (Robinson et al., Citation2014).

Another use (three stories) addresses “wrong” models utilisation for Better understanding the clients’ systems and developing their thinking process. This is a complementary use to other types to create knowledge on non-precise results (Hodges, Citation1991; Hodges & Dewar, Citation1992).

A fifth use (two stories) concerns Promotion or communication on a commercial level, encountering models for sales tool and as evidence of interest when bidding. This use is a direct reference to literature (Hodges, Citation1991; Hodges & Dewar, Citation1992), showing a business-engaging use of “wrong” models in practice.

Lastly, a sixth type is connected to Store and monitor results. This matches storage of information as part of a management system (Hodges, Citation1991; Hodges & Dewar, Citation1992) and suggests that “wrong” models do not need to directly be implemented into actions to offer benefits but instead be saved for later use.

Although the absence of practical studies on uses does not allow comparison with past findings, certain deductions can still be reached. Citation2007 mention how “false“ models may occasionally offer a better fit than “true” models which seems compatible to the current study’s outcomes, while Citation1993, Citation1998) mentions ”exploratory” models as those used for investigating system assumptions and hypotheses regardless of validity. These also lead towards a more complex discussion for future research on model qualities and whether validity should be the main attribute for a model or if usefulness could be adequate. A first consideration may still be reached: the modelling process needs to focus on pragmatic topics and include contents in a model that would deem usefulness to its users – ideally corroborated through model validity. If validity is not achievable, a model could then be considered for uses as the ones presented here.

5.4. “Wrong” models in different domains of application

A last consideration here addresses the potential applicability of “wrong” model uses across different domains and whether these differ.

This study is the first to tackle the idea that certain models may be “wrong“ but still useful. As such it has not focused on separating the results by domain of application. Still, we offer a first indication of whether differences are encountered based on our understanding of the collected data. We notice from that the modellers worked in a range of sectors. We can see from our data that “wrong” models have been used for decisions in manufacturing, and for capacity planning and resource scheduling in services. Similarly, we find ”wrong” models applied to scenario testing in Emergency Departments of clinics.

As such, we have no specific evidence to refute the possibility that a “wrong“ model may be useful in any application domain. This aligns with previous work that explored how simulation could be used in different domains (Robinson, Citation2002b). It appears that the ”mode of practice” employed is more important than the domain of application. For instance, a model used in facilitated workshops (Kotiadis & Tako, Citation2018; Robinson et al., Citation2014; Tako & Kotiadis, Citation2015) requires a lower level of fidelity than a model focused on providing an accurate representation of the real world.

The above also conforms with our previous work which found that simpler simulation models offer similar levels of learning as more complex ones (A. Tako et al., Citation2020). We consider that there is a connection between simple and “wrong“ models in that a simple model could be simplified to such an extent that it could be considered “wrong”. This approach to the definition of wrongness fits with the subjectivity of identifying wrongness in the model. Indeed, our literature analysis has identified the level of detail () as one of the reasons for ”wrong” models (Section 4.1.2). Yet future research could identify potential thresholds of simplification to define model boundaries.

6. Conclusion

This paper is the first empirical study to explore the concept of “wrong” models and their uses.

We conducted interviews with DES modellers offering real-life stories where modelling efforts did not go as planned. This does not mean that “wrong“ models are intentionally developed but rather highlights that the concept of ”wrong” models is subjective and such models may still have some use.

We analysed literature as well as a total of 54 stories on “wrong“ models and defined the main factors leading to the development of such models. We also investigated uses of ”wrong” models in practice, and found that they can be fruitfully used. We recommend that discussions are held between model developers and their users on how best to take advantage of such models.

We also identified concepts that related to “wrong“ models. The authors acknowledge that there is no objective definition of a ”wrong” model as this depends on the interpretation of the model’s validity from the modellers’ or client’s point of view. Our analysis shows that these perceptions may not match and hence highlights the need for reaching a common understanding between clients and modellers. The study also identified the concept of having a useful model, in that a model that offers value to the clients is not the same as that of a valid model.

A number of limitations apply to the study. The subjective categorisation of factors of wrongness and model characteristics is based on the researchers’ viewpoint. Certain characteristics could be considered as more intersectional and under different provided operational definitions. Other researchers could have given different interpretations to the topics and content accounted in the modelling stories and interviews. Also, all stories come from the description of the modellers, meaning that clients’ reactions and inputs were implied based on the modeller’s recollections. The choice of interviewees was subjective and based on availability. A different selection could have led to different outcomes. It is noted that the perceived complexity of each model was subjective and based on each modeller’s description and the researchers’ understanding, without the option for a more objective or quantified interpretation. Due to confidentiality reasons, it was not possible for the researchers to access or analyse these models and their content.

Lastly, the study has identified areas for further analysis. These include considering more specific factors that may have an adverse effect on the perception of “wrong” models and the possibility that different “modes of practice” and domains of application have implications on model “wrongness”. The authors aim to revisit the interviews to explore learning and insights gained from “wrong” models as well as to expand the research to address simplification and usefulness in combination with model wrongness.

Ethical approval

This study has been logged with the School of Business and Economics, Loughborough University, and received ethical clearance on September 20 2018.

Supplemental material

Supplemental Material

Download MS Word (41.9 KB)

Acknowledgments

The authors would like to acknowledge the financial support received from Simul8 Corporation to administer the interviews for this research. Also, we would like to thank the simulation modellers that participated in the interviews and for their input towards this research.

Disclosure statement

No potential conflict of interest was reported by the author(s).

Supplementary material

Supplemental data for this article can be accessed online at https://doi.org/10.1080/17477778.2022.2108736

Notes

1. This research is an extension to the study presented in the Simulation Workshop 2021 (Tsioptsias et al., Citation2021).

2. This last use is known as the “a fortiori” argument (Hodges, Citation1991) and within OR it could refer to a model being used under a new setting to get more reliable insights than those of its initially expected use.

References

  • Balci, O., Nance, R. E., & Arthur, J. D. (2002). Expanding our horizons in verification, validation, and accreditation research and practice. In E. Yücesan, C.-H. Chen, J. L. Snowdon, and J. M. Charnes, eds. Proceedings of the 2002 Winter Simulation Conference, (pp.653–663).
  • Bankes, S. (1993). Exploratory Modeling for policy analysis. Operations Research, 41(3), 435–449. https://doi.org/10.1287/opre.41.3.435
  • Bankes, S. (1998). Policy analysis for complex and uncertain systems through computational experiments. In IEEE Aerospace Conference, vol.1, (pp.9–21).
  • Bazeley, P. (2013). Qualitative data analysis: Practical strategies. SAGE Publications.
  • Box, G. E. P., & Draper, N. R. (1987). Empirical model-building and response surfaces. John Wiley & Sons.
  • Brinkmann, S., & Kvale, S. (2015). Interviews: Learning the craft of qualitative research interviewing (3rd ed.). Sage Publications.
  • Brooks, R. J., & Tobias, A. M. (1996). Choosing the best model: Level of detail, complexity, and model performance. Mathematical and Computer Modelling, 24(4), 1–14. https://doi.org/10.1016/0895-7177(96)00103-3
  • Bryman, A. (2012). Social research methods (4th ed.). Oxford University Press.
  • Castaño, Y. Á. (1999). Dynamic Behavior of NPD. Proceedings of the 17th International Conference of the System Dynamics Society and 5th Australian and New Zealand Systems Conference, (p.10).
  • Chussil, M. (2005). With all this intelligence, why don’t we have better strategies? Journal of Business Strategy, 26(1), 26–33. https://doi.org/10.1108/02756660510575023
  • Eskinasi, M., & Fokkema, E. (2006). Lessons learned from unsuccessful modelling interventions. Systems Research and Behavioral Science, 23(4) , 483–492. https://doi.org/10.1002/sres.774
  • Gass, S. I. (1983). Decision-aiding models: Validation, assessment, and related issues for policy analysis. Operations Research, 31(4), 603–631. https://doi.org/10.1287/opre.31.4.603
  • Gelfert, A. (2019). Assessing the credibility of conceptual models. In C. Beisbart & N. Saam. (Eds.), Computer Simulation Validation. Simulation Foundations, Methods and Applications. (pp. 249–269). Springer. https://doi.org/10.1007/978-3-319-70766-2_10
  • Gogi, A. (2016). Insight generation in simulation studies : an empirical exploration. [ Thesis]. School of Business and Economics, Loughborough University.
  • Goldberg, J., Districh, R., Chen, J. M., Mitwasi, M., Valenzuela, T., & Criss, E. (1990). A simulation model for evaluating a set of emergency vehicle base locations: Development, validation, and usage. Socio-Economic Planning Sciences, 24(2), 125–141. https://doi.org/10.1016/0038-0121(90)90.017-2
  • Harper, A., Mustafee, N., & Yearworth, M. (2021). Facets of trust in simulation studies. European Journal of Operational Research, 289(1), 197–213. https://doi.org/10.1016/j.ejor.2020.06.043
  • Hinkkanen, A., Lang, K. R., & Whinston, A. B. (1995). On the usage of qualitative reasoning as an approach towards enterprise modelling. Annals of Operations Research, 55(1), 101–137. https://doi.org/10.1007/BF02031718
  • Hodges, J. S. (1991). Six (Or So) things you can do with a bad model. Operations Research, 39(3), 355–365. https://doi.org/10.1287/opre.39.3.355
  • Hodges, J., & Dewar, J. (1992). Is it you or your model talking? A framework for model validation. RAND Publication Series.
  • Hooker, J. N. (2007). Good and bad futures for constraint programming (and operations research). Constraint Programming Letters, 1, 21–32. http://public.tepper.cmu.edu/jnh/cpfutures.pdf
  • Jessop, A. (2002). Exploring structure: A block model approach. Civil Engineering and Environmental Systems, 19(4), 263–284. https://doi.org/10.1080/10.286.600.215.048
  • Kasaie, P., Kelton, W. D., Ancona, R., Ward, M. J., Froehle, C. M., & Lyons, M. S. (2017). Lessons learned from the development and parameterization of a computer simulation model to evaluate task modification for health care providers. In Consensus conference, (pp.1–12).
  • Kleijnen, J. P. C. (1995). Statistical validation of simulation models. European Journal of Operational Research, 87(1), 21–34. https://doi.org/10.1016/0377-2217(95)00132-A
  • Kolkman, D. A., Campo, P., Balke-Visser, T., & Gilbert, N. (2016). How to build models for government: Criteria driving model acceptance in policymaking. Policy Sciences, 49(4), 489–504. https://doi.org/10.1007/s11077-016-9250-4
  • Kotiadis, K., & Mingers, J. (2014). Combining PSMs with hard OR methods: The philosophical and practical challenges. In Sally Brailsford, Leonid Churilov, Brian Dangerfield (eds). Discrete‐Event Simulation and System Dynamics for Management Decision Making. (pp. 52–75). John Wiley & Sons, Ltd.
  • Kotiadis, K., Tako, A. A., & Vasilakis, C. (2014). A participative and facilitative conceptual modelling framework for discrete event simulation studies in healthcare. Journal of the Operational Research Society, 65(2), 197–213. https://doi.org/10.1057/jors.2012.176
  • Kotiadis, K., & Tako, A. A. (2018). Facilitated post-model coding in discrete event simulation (DES): A case study in healthcare. European Journal of Operational Research, 266(3), 1120–1133. https://doi.org/10.1016/j.ejor.2017.10.047
  • Lance, C. B., Woehr, D. J., & Meade, A. W. (2007). Case study : A Monte Carlo investigation of assessment center construct validity models. Organizational Research Methods, 10(3), 430–448. https://doi.org/10.1177/1.094.428.106.289.395
  • Law, A. M., & McComas, M. G. (1991).Secrets of successful simulation studies. Proceedings of the 1991 Winter Simulation Conference, (pp.21–27).
  • Lombaers, H. J. M. (1979). Operational research in practice. Engineering and Process Economics, 4(2-3), 245–248. doi:10.1016/0377-841X(79)90036-6.
  • Mens, T., & Van Gorp, P. (2006). A taxonomy of model transformation. Electronic Notes in Theoretical Computer Science, 152(1–2), 125–142. https://doi.org/10.1016/j.entcs.2005.10.021
  • Mingers, J. (2003). A classification of the philosophical assumptions of management sciencemethods. Journal of the Operational Research Society, 54(6), 559–570. https://doi.org/10.1057/palgrave.jors.2601436
  • Ormerod, J. (2018). The logic and methods of OR consulting practice: Towards a foundational view. Journal of the Operational Research Society, 69(9), 1357–1378. https://doi.org/10.1080/01605682.2017.1392407
  • Phillips, L. D. (1984). A theory of requisite decision models. Acta Psychologica, 56 (1-3), 29–48. https://doi.org/10.1016/0001-6918(84)90,005-2
  • Pidd, M. (2004). Systems modelling theory and practice. John Wiley & Sons.
  • Pidd, M. (2009). Tools for thinking: Modelling in management science (3rd ed.). John Wiley and Sons Ltd.
  • Prieto, D. M., Das, T. K., Savachkin, A. A., Uribe, A., Izurieta, R., & Malavade, S. (2012). A systematic review to identify areas of enhancements of pandemic simulation models for operational use at provincial and local levels. BMC Public Health, 12(1), 1–13. https://doi.org/10.1186/1471-2458-12-251
  • Robinson, S., & Pidd, M. (1998). Provider and customer expectations of successful simulation projects. Journal of the Operational Research Society, 49(3), 200–209. https://doi.org/10.1057/palgrave.jors.2600516
  • Robinson, S. (1999). Three sources of simulation inaccuracy (and how to overcome them). In P. A. Farrington, H. B. Nembhard, D. T. Sturrock, and G. W. Evans (Eds), Proceedings of the 1999 Winter Simulation Conference. (pp.1701–1708).
  • Robinson, S. (2002a). General concepts of quality for discrete-event simulation. European Journal of Operational Research, 138(1), 103–117. https://doi.org/10.1016/S0377-2217(01)00127-8
  • Robinson, S. (2002b). Modes of simulation practice: Approaches to business and military simulation. Simulation Modelling Practice and Theory, 10(8), 513–523. https://doi.org/10.1016/S1569-190X(02)00117-X
  • Robinson, S. (2014). Simulation - the practice of model development and use (2nd ed.). Macmillan education, Palgrave.
  • Robinson, S., Worthington, C., Burgess, N., & Radnor, Z. J. (2014). Facilitated modelling with discrete-event simulation: Reality or myth? European Journal of Operational Research, 234(1), 231–240. https://doi.org/10.1016/j.ejor.2012.12.024
  • Robinson, S. (2015). A tutorial on conceptual modeling for simulation. In L. Yilmaz, W. K. V. Chan, I. Moon, T. M. K. Roeder, C. Macal, and M. D. Rossetti (Eds), Proceedings of the 2015 Winter Simulation Conference, (pp.1820–1834).
  • Rouwette, E. A. J. A., Vennix, J. A. M., & Van Mullekom, T. (2002). Group model building effectiveness: A review of assessment studies. System Dynamics Review, 18(1), 5–45. https://doi.org/10.1002/sdr.229
  • Salt, J. D. (2008). The seven habits of highly defective simulation projects. Journal of Simulation, 2(3), 155–161. https://doi.org/10.1057/jos.2008.7
  • Sargent, R. G. (1991). Simulation model verification and validation. In B. L. Nelson, W. D. Kelton, & G. M. Clark, eds., Proceedings of the 23rd conference on Winter simulation, (pp.37–47).
  • Tako, A. A., Kotiadis, K., & Vasilakis, C. (2010). A participative modelling framework for developing conceptual models in healthcare simulation studies. In Proceedings of the 2010 Winter Simulation Conference, (pp. 500–512).
  • Tako, A., & Kotiadis, K. (2015). PartiSim: A multi-methodology framework to support facilitated simulation modelling in healthcare. European Journal of Operational Research, 244(2), 555–564. https://doi.org/10.1016/j.ejor.2015.01.046
  • Tako, A. A. (2015). Exploring the model development process in discrete-event simulation: Insights from six expert modellers. Journal of the Operational Research Society, 66(5), 747–760. https://doi.org/10.1057/jors.2014.52
  • Tako, A., Tsioptsias, N., & Robinson, S. (2020). Can we learning from simplified simulation models? An experimental study on user learning. Journal Of Simulation, 14(2), 130–144. https://doi.org/10.1080/17.477.778.2019.1704636
  • Tolk, A. (2015). Learning something right from models that are wrong: Epistemology of simulation. In L. InYilmaz (Ed.), Concepts and Methodologies for Modeling and Simulation. Simulation Foundations, Methods andApplications. Springer. (pp. 87–106). https://doi.org/10.1007/978-3-319-15096-3_5
  • Tsioptsias, N., Tako, A., & Robinson, S. (2021). An exploratory study on the uses of “wrong” discrete event simulation models in practice. In Proceedings of the Operational Research Society Simulation Workshop 2021 (SW21), (pp.168–177).
  • Vennix, J. A. (1999). Group model-building: Tackling messy problems. System Dynamics Review, 15(4), 379–401. https://doi.org/10.1002/(SICI)1099-1727(199.924)15:4<379::AID-SDR179>3.0.CO;2-E
  • Williams, H. P. (1989). How important are models to operational research? IMA Journal of Management Mathematics, 2(2), 189–195. https://doi.org/10.1093/imaman/2.2.189
  • Yilmaz, L., & Liu, B. (2020). Model credibility revisited: Concepts and considerations for appropriate trust. Journal of Simulation. https://doi.org/10.1080/17477778.2020.1821587