2,703
Views
2
CrossRef citations to date
0
Altmetric
Research Articles

The Patient is Thriving! Current Issues, Recent Advances, and Future Directions in Creativity Assessment

Pages 291-303 | Received 20 Jun 2022, Published online: 18 Aug 2022

ABSTRACT

In 1998, Plucker and Runco provided an overview of creativity assessment, noting current issues (fluency confounds, generality vs. specificity), recent advances (predictive validity, implicit theories), and promising future directions (moving beyond divergent thinking measures, reliance on batteries of assessments, translation into practice). In the ensuing quarter century, the field experienced large growth in the quantity, breadth, and depth of assessment work, suggesting another analysis is timely. The purpose of this paper is to review the 1998 analysis and identify current issues, advances, and future directions for creativity measurement. Recent advances include growth in assessment quantity and quality and use of semantic distance as a scoring technique. Current issues include mismatches between current conceptions of creativity and those on which many measures are based, the need for psychometric quality standards, and a paucity of predictive validity evidence. The paper concludes with analysis of likely future directions, including use of machine learning to administer and score assessments and refinement of our conceptual frameworks for creativity assessment. Although the 1998 paper was written within an academic climate of harsh criticism of creativity measurement, the current climate is more positive, with reason for optimism about the continued growth of this important aspect of the field.

Across a number of fields, identifying and measuring creativity continues to grow in prevalence and importance. Within the private sector, businesses are focusing considerable resources on talent development, often with an emphasis on innovative skills. K-12 educators and policymakers seek ways to assess creativity to provide an alternative set of student outcomes that supplement traditional high-stakes testing. Many colleges and universities around the world have refocused their programs to help students develop innovative and entrepreneurial skills. And, of course, researchers across a number of fields rely on measures of creativity in their research and evaluation work.

Shortly before the turn of the previous century, many scholars noted that creativity assessment was under siege, with strident criticisms from both within (Baer, Citation1994; Cropley, Citation2000; Hunsaker & Callahan, Citation1995; Plucker & Renzulli, Citation1999) and outside the field (Eysenck, Citation1994; Houtz & Krug, Citation1995). Criticisms included a perceived lack of predictive validity evidence, a lack of content-specific measures at a time where context was being emphasized in education theory and research, overreliance on classical test theory and older measurement models, and an overreliance on divergent thinking tasks for assessment of creativity and creative potential. Supporters of assessment research granted the critics some of these points but noted that the critiques went too far or were simply incorrect.

Within this context, Plucker and Runco (Citation1998) undertook a comprehensive review of creativity measurement, including both controversies and possible future directions. The paper was titled, “The death of creativity measurement has been greatly exaggerated,” because most public commentary on assessment of creativity had become harshly negative. Yet the paper’s overall conclusion was that important, recent advances had occurred in several aspects of creativity assessment, making them optimistic about the future of this aspect of the field.

A quarter century later, this optimism appears to have been warranted. There has been exponential growth in research and development on measurement related to creativity, entrepreneurship, and innovation, with several important lines of research on improvement of existing measures and creation of new assessment strategies. For this reason, revisiting the previous analysis is timely, as it provides the field with an opportunity to take stock of its recent progress, identify current and likely future trends, and note areas in need of additional effort.Footnote1

From 1998 to now

Plucker and Runco (Citation1998) offered three sets of observations: current issues being debated in the field, recent advances in creativity assessment, and future directions that could prove helpful to creativity measurement and the field in general. This section notes the extent to which subsequent research addressed the issues, continued or moved away from the trends, and matched the predictions.

Have we made progress on the field’s big assessment issues?

Two issues were noted as current issues in 1998, debates over domain generality and specificity and fluency as a contaminating factor in scoring of divergent thinking tasks. Regarding specificity-generality, many fields were incorporating sociocultural perspectives, with their emphasis on context and person-environment interaction (e.g., Bredo, Citation1994; O’Loughlin, Citation1992), at the turn of the century, and creativity scholarship was not an exception. Reacting to a strong domain-general view for much of the field’s history, Baer (Citation1998) and others noted the need for strong domain specific theories and models with accompanying assessments.

The field’s emphasis on domain specificity became quite strong in ensuing years, but that pendulum has swung back toward the middle (again, as in many other fields). Although the evidence of content generality has been the subject of several recent studies (Qian & Plucker, Citation2018; Qian, Plucker, & Yang, Citation2019; van Broekhoven, Cropley, & Seegers, Citation2020), most models embrace both domain specific and domain general aspects of creativity. For example, the amusement park model of creativity proposes that creativity becomes more domain-specific as one moves from initial stages of the creative process to application in domains and micro-domains (Baer & Kaufman, Citation2005), and Plucker and Beghetto (Citation2004) offered a conceptualization that emphasized the importance between a balance of domain general and domain specific attitudes, skills, and knowledge, with extremes in either direction likely hindering creativity (i.e., too much specificity leading to functional fixedness, and too much generality leading to superficiality).

Although this debate does not yet appear to be settled (and may never be!), scholars continue to put serious thought into the role of context and creativity. Regarding implications for assessment, Plucker, Meyer, and Makel (Citationin press) suggest a developmental perspective: Assessing under an assumption of domain and task generality is reasonable for younger students, but domain specificity is appropriate for older students and adults, given the importance of the application of creativity to specific tasks and using specific content as one ages and moves into the workforce.

The idea of fluency of a contaminating factor in the scoring of performance on divergent thinking tasks goes back to at least Clark and Mirels (Citation1970). They noted the high intercorrelations among creativity and intelligence measures, hypothesizing that fluency was the cause of these correlations. Intercorrelations dropped in magnitude considerably when fluency was controlled. Subsequently, several other researchers also noted the potentially confounding effects of fluency when scoring assessments, with several offering strategies for controlling for fluency effects (e.g., Feldhusen & Goh, Citation1995; Hocevar, Citation1979; Runco & Albert, Citation1985). Plucker and Runco (Citation1998) recommended that the various strategies be examined in light of their impact on reliability and validity evidence, not dissimilar from the recent call by Barbot, Hass, and Reiter-Palmon (Citation2019) for homogenization of measurement standards.

Although researchers attempted to address the role of fluency over the past couple decades (e.g., Plucker, Qian, & Schmalensee, Citation2014), the field moved far beyond these recommendations and undertook a wide-ranging analysis of new or improved scoring methods for DT task performance. Much of this work built on earlier efforts by Runco and colleagues (e.g., Runco & Mraz, Citation1992; Runco, Okuda, & Thurston, Citation1987), which was built upon by Silvia and colleagues in a number of important studies in the early 2000s (Silvia, Citation2011; Silvia, Martin, & Nusbaum, Citation2009; Silvia et al., Citation2008). Since then, the amount of DT scoring work has grown exponentially. Acar and Runco (Citation2019) provide a comprehensive overview of the large number of studies on this topic.

Collectively, the sharp growth in DT scoring research is a very positive development for the field relative to a focus on fluency effects, given that (a) DT scoring in general needed fresh perspectives and (b) evidence suggests that fluency may have a genetic basis (i.e., it may be important in its own right and not “controlled for”; see Runco et al., Citation2011). Notably, Reiter-Palmon, Forthmann, and Barbot (Citation2019) have proposed a framework for classifying the various aspects of scoring that need to be considered and reported in studies using DT-based assessments, including objectivity vs. subjectivity, individual item scoring vs. item pool scoring, and scoring dimension (fluency, originality, etc.).

Did recent advances continue to advance?

Two recent advances were studies of the predictive validity of assessments and increased sophistication in the study of implicit theories. Plucker and Runco noted that predictive validity of DT measures was generally assessed to be unimpressive (Baer, Citation1993; Wallach, Citation1976), but that the available evidence was more promising than many critics suggested, especially for domain-specific DT tasks (Hong, Milgram, & Gorsky, Citation1995; Okuda, Runco, & Berger, Citation1991; Sawyers & Canestaro, Citation1989). At the same time, they called for more such research, noting its importance to our understanding of creativity assessment and creative development across the lifespan.

Unfortunately, the ensuing quarter-century has seen few such studies. Torrance’s longitudinal study, which first collected data in 1958, has been reanalyzed and extended several times, with similarly positive conclusions about the predictive validity of the DT measures. For example, Plucker’s (Citation1999) reanalysis of the Torrance data suggests childhood creativity predicted adult creative accomplishment, and that it did so at a higher level than intelligence test scores. Runco, Millar, Acar, and Cramond (Citation2010) subsequently collected new data from the subjects in that study and found a positive if more complex pattern of longitudinal relationships. A meta-analysis of DT scores from students in grades 1–12 provides evidence that performance generally increases over time, but this study speaks more to development than predictive validity (Said-Metwaly, Fernández-Castilla, Kyndt, Van den Noortgate, & Barbot, Citation2021).

However, the predictive validity of other forms of creativity assessment is very limited. Kaufman et al. (Citation2008) noted that critics were holding DT assessments to a predictive validity standard that was not being applied equally to other popular assessments (e.g., the consensual assessment technique). This situation has improved little since 1998 and 2008.

Implicit theories were a popular area of scholarship, with considerable innovation into the late 1990s on assessment of those theories and personal beliefs. This trend continued into the early 2000s (Puccio & Chimento, Citation2001), but the momentum appeared to slow considerably before increasing again in the mid-2010s. The type of implicit theory study also appears to have changed over the past two decades. In the earlier body of work, the majority of studies focused on differences in implicit theories of creativity among cultures (Chan & Chan, Citation1999; Lim & Plucker, Citation2001; Niu & Sternberg, Citation2002; Paletz & Peng, Citation2008; Panda & Yadava, Citation2005). Recent work has focused on implicit theories held by specific groups, such as teachers, professors, and students (Cropley, Patston, Marrone, & Kaufman, Citation2019; Gralewski & Karwowski, Citation2018; Maksić & Spasenović, Citation2018; Pavlović & Maksić, Citation2019). However, there are plentiful exceptions to these trends (e.g., Loewenstein & Mueller, Citation2016; Patston, Cropley, Marrone, & Kaufman, Citation2018; Spiel & von Korff, Citation1998).

However, the increase in quantity and sophistication of implicit theory studies does not appear to be matched with parallel development in the sophistication of assessment of those theories. For example, Malmelin and Nivari-Lindström (Citation2017) assessed implicit theories by having journalists respond to a series of statements about the nature of creativity, which they then analyzed qualitatively to arrive at an understanding of how Finnish journalists conceptualize creative work. The study is unique in its highly applied, contextualized focus on professional journalists and its attention to details such as the use of “creative work” rather than “creativity” (see Runco, Citation2007). But the actual assessment strategy is very straightforward and not distinct from strategies used to assess implicit theories in the mid-1980s. Similarly, but using a more traditional quantitative approach for data analysis, Cropley et al. (Citation2019) asked teachers to respond to a series of statements about creativity (e.g., engaging students’ creativity helps their learning, it is possible to measure student creativity, creative students are the ones who write, draw, or play music), with several items reverse-coded. They found support for five factors of teacher beliefs, namely that creativity is important for education, requires domain-specific knowledge, can be measured, is found in all subjects, and is not limited to certain students.

The recent research of implicit theories is interesting and help to build our knowledge base of how various groups of people think about creativity. But it is somewhat surprising that developments from the 1980s and 1990s regarding the assessment of implicit theories have not been incorporated or improved upon in the contemporary studies. For example, researchers from that earlier period found evidence that responses to close- and open-ended questions about creativity beliefs tended to reflect social desirability, but that when asked to use their implicit theories (such as in the evaluation of profiles of potentially creative students or employees), they appear to use different, more pragmatic conceptions of creativity (Lim & Plucker, Citation2001; Sternberg, Citation1985). The majority of recent studies on this topic appear to use only participant self-responses. The assessment of implicit theories of creativity appears to have become much broader regarding types of participants and their roles; but it has not become deeper over the past quarter century and may, in fact, have become less sophisticated.

Were the future directions fruitful?

The 1998 analysis also recommended three future directions for creativity assessment: broadened application of psychometric methods, reliance on batteries of assessments, and translation into practice. Regarding psychometric methods, few published studies were conducted from the perspective of modern measurement theories and techniques, such as item response theory, and the majority of studies did not use more sophisticated statistical methods, such as structural equation modeling. The field has moved in this direction, with more sophisticated, contemporary measurement strategies (e.g., Becker, Cabeza, Shaw, & Davidson, Citation2022; Primi, Silvia, Jauk, & Benedek, Citation2019; Qian & Plucker, Citation2018; Van Hooijdonk, Mainhard, Kroesbergen, & Van Tartwijk, Citation2022; Wang & Long, Citation2022). This work includes assessment of group creativity (Han, Long, Ge, & Pang, Citation2022; Hester & Hester, Citation2012; Romero et al., Citation2019), which is also being pursued by the Schmidt Futures Foundation.

This sophistication is also evident in conceptual analyses (Cerrato, Siano, Marco, & Ricci, Citation2019; Kaufman, Citation2019), such as reviews of research on the consensual assessment technique (Cseh & Jefferies, Citation2019; Myszkowski & Storme, Citation2019). In particular, Myszkowski and Storme propose the use of item response theory models to improve the CAT’s psychometric framework. Continuing to apply advances in measurement theory and technique will similarly advance development of the field’s creativity measures.

The growth of systems and sociocultural approaches to creativity led Plucker and Runco to recommend that creativity researchers use multiple types of measures when studying creativity.Footnote2 There is little evidence of reliance on batteries of assessments in creativity studies over the past 25 years. For example, in the most recent issues of three major journals for the scientific study of creativity, 11 empirical articles included the assessment of creativity. Only six relied on more than one type of measure, with only one using more than two distinct assessment types. The exception here proves the rule: Ceh, Edelmann, Hofer, and Benedek (Citation2022), in an examination of factors that correlate with novice raters’ ability to discern more and less creative responses to DT tasks, used divergent thinking tasks, personality measures, an activity checklist, and evaluation of others’ creativity. The authors conclude that DT performance, past creative activities, and openness to experience (among other factors) predict discernment. By using a broad and rather comprehensive set of assessment types, the researchers were able to provide a more useful set of insights than if they had used just one or two assessment types. Barbot et al. (Citation2019) note that particular types of creativity assessments “ … are all suboptimal, in that they incompletely represent the broader phenomenon” (p. 238). Using batteries of diverse assessments is the best path forward to a comprehensive understanding of our field’s complex construct.

Plucker and Runco noted that “Creativity researchers face one particular problem that is not always shared with other investigators of psychological phenomenon: The need to translate our work into practice” (p. 38). They noted recent application of creativity assessment strategies to a number of fields (e.g., Clements, Citation1991; Cole, Sugioka, & Yamagata-Lynch, Citation1999; Couger & Dengate, Citation1992; Gilbert & Prenshaw, Citation1996) and recommended that researchers continue to develop measures that could be easily used in applied settings. The growth of applied measurement of creativity has occurred in a wide range of fields, including athletics (Fardilha & Allen, Citation2020), writing (D’Souza, Citation2021), science education (Li, Cai, Kuznetsova, & Kurilova, Citation2022; Wang & Long, Citation2022), design (Dong, Zhu, & Li, Citation2021; Han, Forbes, & Schaefer, Citation2019), higher education (Snyder, Hammond, Grohman, & Katz-Buonincontro, Citation2019), and computing education (Lehmkuhl, Gresse von Wangenheim, Martins-Pacheco, Borgatto, & da Cruz Alves, Citation2021), among many other fields. Some of these lines of research and development are quite complex, such as Wang and Long’s application of Rasch models to the analysis of consensual assessment technique scores on science creativity tasks. This “future direction” has largely been realized.

But that is not to say that additional applied work is unnecessary, as several important topics remain to be addressed. For example, questions about the impact of rubrics on creativity have been consistently raised for years, with almost no empirical research to guide practice. Given that assessment rubrics are all the rage in K-12 and higher education – and often recommended as a best practice – the lack of research is glaring. As a case in point, the most comprehensive and thoughtful review on the use of rubrics in higher education does not mention creativity (Dawson, Citation2017). The goal of this research need not be a simple thumbs-up/thumbs-down on rubrics, but rather a more nuanced examination of how rubrics can be used to best support creativity (e.g., Do single-point rubrics provide the necessary balance of guidance and flexibility to optimize both student learning and creativity? See Fluckiger, Citation2010). Similarly, with the emphasis on group work in both education and the private sector, assessment approaches that allow for simultaneous monitoring of both individual and group creativity within groups are needed (see Stefanic & Randles, Citation2015).

Current state of creativity assessment

Given the mixed progress in the areas identified in 1998, what are the current issues and needed future directions of the field as its researchers work in the third decade of the 21st century? includes a summary of the 1998 themes and 2022 issues, advances, and potential future directions. In this section, the 2022 trends are discussed in more detail.Footnote3 The new categories overlap several of the old observations, and the lack of mention of implicit theories, for instance, should not be viewed as that work no longer being important, but rather subsumed by the broader category of continuing to improve assessment quality. More to the point, the lack of a specific category related to assessment batteries is due more to the inclusion of batteries across a number of the new categories rather than a deemphasis of their importance.

Table 1. Summary of current issues, recent advances, and future directions in creativity assessment, 1998 vs. 2022.

Recent advances

By far the biggest change since 1998 is the sharp increase in assessment quantity and quality. Every major research journal that specializes in creativity research routinely publishes assessment articles, and the 2019 special assessment issue of Psychology of Aesthetics, Creativity, and the Arts is a testament to the hundreds of studies that have been published over the past quarter century. Regarding quantity, dozens of new assessments have appeared since 2000, representing a wide range of creativity dimensions. For example, the Kaufman Domains of Creativity scales are activity checklists in the domains of everyday, scholarly, performance (writing and music), scientific, and artistic creativity (Kaufman, Citation2012). Not only have the K-DOC scales been the subject of several measurement studies (e.g., Kapoor, Reiter-Palmon, & Kaufman, Citation2021; McKay, Karwowski, & Kaufman, Citation2017), the scales have been used as criterion or independent variables in several subsequent studies of creativity, including studies of creative self-efficacy (Barbot, Citation2020), problem-solving in science (Aschauer, Haim, & Weber, Citation2021), and musical creativity (Tavani, Caroff, Storme, & Collange, Citation2016).

Similarly, the Runco Ideational Behavior Scale (Runco et al., Citation2001) was developed as a criterion measure for creativity studies, with the logic that attitudes toward ideational behavior were often the desired outcome of interest for many intervention studies. The scale has been examined in several measurement studies (e.g., Runco et al., Citation2014; Sen, Citation2016, Citation2022) and used as a criterion variable in studies of creativity (Paek, Park, Runco, & Choe, Citation2016; Plucker et al., Citation2006). Almost countless other measures have emerged this century, many with substantial psychometric support and in wide use, such as the Biographical Inventory of Creative Behaviors (BICB; Silvia et al., Citation2021), the Creative Achievement Questionnaire (CAQ; Carson, Peterson, & Higgins, Citation2005), and variations of the divergent thinking tasks mentioned earlier. The sheer quantity of new and/or improved assessments for creativity is staggering.

At the same time, the quality of these new measures is also impressive. As noted earlier, measurement studies and critical reviews have sharply increased in prevalence (e.g., Kaufman, Citation2019; Silvia et al., Citation2012), and many use modern methods such as Rasch modeling. The Consensual Assessment Technique, befitting its popularity, has also been the subject of extensive measurement studies that are helping researchers improve its use (see Cseh & Jefferies, Citation2019). The plethora of studies on administration and scoring of divergent thinking tasks is leading to considerable insights on how best to use this popular assessment strategy. Plucker and Makel (Citation2010) wryly noted that users of creativity assessments needed to be patient regarding psychometric integrity, given that developers of creativity measures had not had the decades of experience and funding enjoyed by developers of achievement, ability, and intelligence tests. That patience has clearly been rewarded with a wide range of high-quality assessments.

This positive evaluation of the field’s progress comes with an important caveat: There appears to be a trend toward creating new instruments when there are already several quality instruments that are in regular use. These new instruments appear to provide incremental improvements when their authors believe them to be big leaps in new directions. The 10th new behavior checklist will likely have less future value to the field than the development of a qualitatively different approach to assessment of mathematical creativity (Tan, Mourgues, Bolden, & Grigorenko, Citation2014) or significant improvements in technologies that allow traditionally hand-scored measures to be automated and scaled. Being creative about new forms of creativity assessment would serve the field and its stakeholders well.

Another important advance is the estimation of semantic distance. In brief, semantic distance is an estimate of concept dissimilarity; given that the scoring of originality in verbal DT tasks is often based on dissimilarity, the application semantic distance to that class of creativity assessments has become quite popular. In particular, this approach to analyzing DT task responses (primarily latent semantic analysis or LSA) is both theoretically grounded (Acar & Runco, Citation2019; Kenett, Citation2019) and empirically promising (Beaty & Johnson, Citation2021; Heinen & Johnson, Citation2018; Orwig, Diez, Vannini, Beaty, & Sepulcre, Citation2021). These methods, which decrease scoring time due to automation, are not without their critics, who note that the generalizability of LSA to other DT tasks may be limited (see Barbot et al., Citation2019). Regardless, these techniques may have broader applications (e.g., Guo, Ge, & Pang, Citation2019).

Current issues

A current issue that the 1998 analysis did not anticipate was conceptual, although it has significant implications for creativity assessment: How do we define creativity? Or more to the point, what are we assessing when we say we are assessing “creativity?” Considerable effort has been devoted to elucidating our definitions, with general agreement that originality and utility are key components (Runco & Jaeger, Citation2012). There is some diversity of opinion about extensions of that standard definition, with Simonton (Citation2012) borrowing from the U.S. Patent Office definition – proposing surprising, and Plucker, Beghetto, and Dow (Citation2004) rooting the definition within specific social contexts. Although some of the field’s leaders see these explorations as unfruitful (e.g., Silvia, Citation2018), these conceptual discussions have helped unify the field. To put it personally, in the early 2000s I attended a conference on creativity organized by the White House’s Office of Science Technology and Policy that was sponsored by the leading U.S. government research agencies and major private foundations. The speakers agreed on one thing: Creativity couldn’t be defined or, therefore, assessed. That was a common attitude outside the field. Today these experiences are rare, in part because the literature is replete with well-defined constructs, and researchers have disseminated their conceptual and assessment work broadly. This development is a major credit to the field, one that will benefit researchers and stakeholders for many years to come.

At the same time, two additional conceptual issues have emerged. The first deals with the broad models of creativity that the field produces. Given the wide-ranging nature of the construct, with its many personal, cognitive, and social dimensions, building robust frameworks for classifying these dimensions will be of great utility to those working with creativity assessments. As a case in point, Guilford’s (Citation1967) original version of the Structure-of-the-Intellect model included 24 distinct types of divergent production, a number that expanded with the growth of the SOI model. Recently, Acar and Runco (Citation2019) have proposed a new organizing framework for divergent thinking, Reiter Palmon et al. (Citation2019) provided a detailed framework for divergent thinking assessment dimensions, and Kaufman (Citation2019) proposed a classification of self-report measures. At this point in the field’s development – at least from an assessment perspective – further development of organizing frameworks is likely more helpful than additional debates over broad definitions, the majority of which end up being quite similar and only rhetorically distinct.

Second, sociocultural perspectives on creativity and talent development continue to gain influence within the field. These conceptualizations place a strong emphasis on the interaction among the individual and their social context (Glăveanu, Citation2015; Glăveanu et al., Citation2019; Simonton, Citation2019), with some scholars proposing that the individual creator is inseparable from their environment and social interactions (Plucker, McWilliams, & Guo, Citation2021). These perspectives have implications for creativity assessment that are just beginning to be explored and should be a focus of future creativity efforts. Measurement approaches that focus narrowly on specific aspects of creativity have less utility from this philosophical perspective, again making the use of batteries of assessments appealing in creativity research.

Another critical issue is the need for quality standards for the field’s assessments. Barbot et al. (Citation2019) propose two broad criteria: transparency and homogenization. Transparent reporting practices are important because researchers should ensure “the particular facet(s) of the creativity phenomenon being investigated are themselves clearly defined and operationalized” (p. 238) and important but hardly new idea, both within our field (see Plucker et al., Citation2004) and scholarship in general. If we are not clear about what we think we are studying and how we are choosing to study it, interpretation of our research base is at best difficult and at worst nearly impossible.

Regarding homogenization, Barbot et al. call for standardization of measurement “at the subconstruct or facet level” (p. 238), meaning that specific aspects of creativity, such as DT, should have agreed-upon measures, administration protocols, and scoring strategies. Essentially, if transparency helps a research consumer understand how three studies on the same topic within creativity may differ, homogenization addresses the fact that perhaps those three research teams should be using similar methods. For example, if three groups conduct studies of the relationship between creative self-efficacy and divergent thinking within the context of math problem solving, transparency would help readers understand how each team measured and scored self-efficacy and DT differently. Homogenization would encourage the research teams to use assessment strategies that were similar if not identical, facilitating generalization of results across studies. The transparency + homogenization approach should serve as a foundation for future efforts to ensure quality and progress in a field that is, by definition, prone to divergence.

A possible, third criterion is context and inclusion. Given the growing emphasis on sociocultural perspectives in the field, assumptions about generalization across contexts – especially cultural contexts – need to be challenged. A major problem within the social sciences has been an overreliance on the use of WEIRD (Western, Educated, Industrialized, Rich, and Democratic) samples. Creativity research appears to be better than other fields at using non-WEIRD samples, which is a positive development. For example, many of the K-DOC and RIBS measurement studies have involved non-WEIRD samples (Awofala & Fatade, Citation2015; Kalis & Roke, Citation2011; Kandemir & Kaufman, Citation2020; Susanto, Novitasari, Rakhmat, Hidayat, & Wibowo, Citation2018; Tep, Maneewan, & Chuathong, Citation2021; Tsai, Citation2015), providing evidence of the impact of cultural context on these particular instruments and their resulting data. This criterion is related to transparency and homogenization but is distinct enough (and important enough) that it can stand on its own.

Although an elaborated Barbot et al. framework is a good foundation for future assessment efforts within the field, it should probably be treated as aspirational. For example, the transparency criterion was emphasized by Plucker et al. (Citation2004), yet a follow-up study by Puryear and Lamb (2020) found only marginal improvement in reporting practices. Of course, journal editors could have a major impact on reporting practices, much as they have had a major impact on open science practices more broadly. By requiring authors of all submitted manuscripts to, at the very least, define their terms and constructs, transparency in the field’s assessment research (and research in general) would leap forward.

That said, homogenization of assessment strategies feels unlikely. Given the often heated competitiveness among researchers in the field, it is not unreasonable to question the degree to which we could collectively agree on standard approaches on even the finest-grained assessments. Given that each major research team around the world has developed their own creative behavior checklist (a small exaggeration only), is there any hope of agreement on assessment of a more complex subconstruct such as divergent thinking? Not to put too fine a point on it, but after more than a quarter century of use of the CAT, why are we no closer to knowing the optimal characteristics of raters? Again, progress toward homogenization is a desirable aspiration, but that progress will almost certainly be measured in baby steps.

A third critical issue is the need for more predictive validity studies of creativity assessment. Although the nature of validity has evolved over the past few decades (see Kane, Citation2016), evidence that assessment scores predict future behavior and other desirable outcomes remains important. As noted earlier, the field has made little progress in this respect and needs additional work in this area. Even short-term studies of two-to-three years would provide insights that facilitate improvement of assessment instruments, models, and strategies.

The importance of predictive validity is often debated (i.e., I have been told by more than one proponent of the CAT that predictive validity evidence is unnecessary). But our measures, especially when used in applied research, often carry the implicit promise that the data being collected is important for the future. For example, personality measures of creativity are often used in workplace hiring, and cognitive measures are used in schools. If these and other assessment strategies truly do not have any predictive power, the usefulness of creativity in general will rightfully come under scrutiny. The limited evidence is that many of the most popular measures are associated with evidence of predictive validity, providing reason for optimism. But a little more data would make the need for optimism less necessary.

Future directions

The incorporation of machine learning in the scoring of creativity assessments can certainly be classified as a “recent advance,” given the breadth and depth of ongoing efforts (see earlier discussion about semantic distance scoring). Yet the application of artificial intelligence (AI) looks to be such an important part of the field moving forward that it is best categorized as a future direction.

The use of AI in creativity assessment is particular exciting because it will allow us to do what we already do better while also allowing us to do new things. For example, the inability to scale creativity measures due to scoring burden is a major impediment to the inclusion of such measures in K-12 accountability systems and similar, wider applications in education and the private sector (Plucker & Alanazi, Citation2019). At the same time, unique innovations (e.g., use of semantic distance) were difficult to tackle given the required computing power and related resources. AI may help researchers address both issues by allowing us to scale things we already do well while opening doors to new possibilities that simply were not attainable in an analog world.

The field is moving assertively in this direction, with several teams around the world developing assessments (and, to a lesser extent, assessment systems) that rely on or at least use some aspects of AI (e.g., Beaty & Johnson, Citation2021; Dumas, Organisciak, & Doherty, Citation2021; Kovalkov, Paaßen, Segal, Pinkwart, & Gal, Citation2021; Sung, Cheng, Tseng, Chang, & Lin, Citation2022). These efforts predominantly focus on written language, although efforts involving figural creativity and speech are also ongoing (e.g., Cerrato et al., Citation2019; Cropley & Marrone, Citation2021; Plucker et al., Citationin press; Shute & Rahimi, Citation2021). The efforts focused on written input feel like an example of the first advantage (i.e., doing what we already do better and at greater scale), and the non-written initiatives feel like the second advantage (opening ourselves up to innovative ways to assess creativity), although there is clearly overlap.

In addition, an exciting but rarely mentioned benefit of these projects is the sheer number of computer scientists, engineers, and other non-usual suspects contributing to our field. Much as the study of creativity benefited immensely from the influx of creativity researchers from East Asia and Central and Eastern Europe over the past 20 years, the growth in non-psychologists contributing to the field, in combination with continued growth of creativity research around the globe, promises to continue increasing the quantity and quality of creativity scholarship in general and creativity assessment in particular.

Although the use of modern computing power to improve creativity assessment bodes well for the future, researchers should continue to focus on conceptual issues (e.g., Basalla, Schneider, & Vom Brocke, Citation2022). With technology allowing us to do more and different things, greatly expanding the possibilities for creativity assessment, serious consideration of why and how we assess creativity in certain contexts will become increasingly important. Put differently, advances in the coming decade may allow us to assess creativity in almost any setting with previously unattainable speed. The possibilities are almost endless, but they are not equally valuable. What are the most promising approaches to large-scale creativity assessment, and which are the most valuable for improving human performance and quality of life?

For example, Plucker et al. (Citationin press) have noted that the multifaceted, sociocultural, and developmental natures of creativity make it imperative to adapt assessments for different life stages and problem-solving contexts (e.g., PreK-12, postsecondary, workforce). They propose a developmental framework to serve as a guide to future development of creativity assessment systems, suggesting, for example, that domain-general assessments are appropriate for young children, domain-specific measures become more important for high school and college students, and domain- and even task-specific measures are most relevant when working with adults in the workforce. Plucker et al. also note that much of the AI-assisted work in the field focuses on cognition and process, which are an important aspect of creativity but not the only important aspect. Conceptual work on creativity assessment design and implementation may not be as sexy as developing machine language models, but it is necessary for the field to move efficiently in positive directions (see also Vartanian, Citation2014).

Conclusion

It is difficult to transport oneself back in time 25 years to answer the question, “If you knew what was going to happen over these next few decades, would you feel like progress was made?” Based on the analysis above, with progress in some areas but not others (and even a little backsliding), it is tempting to conclude that the authors of the 1998 paper would deliver a mixed verdict. But more likely, I suspect we would have been thrilled. The assessment of creativity has never been more dynamic, due to the tremendous volume of such work, the quality and technical sophistication of that work, and the diversity and number of researchers conducting it. In fact, as the original paper was written in the mid- to late-90s, it would have been difficult if not impossible to envision that the assessment of creativity – and the science of creativity in general – could become this robust. Much important work remains to be done, but it will be entertaining and fulfilling to see the new directions the field takes over the next quarter-century.

Disclosure statement

No potential conflict of interest was reported by the author(s).

Notes

1. The sheer volume of creativity assessment scholarship over the past 25 years makes a truly comprehensive citation of relevant work impractical – the reference list alone would be several dozen pages long. The goal of this paper is to highlight key studies and representative work and not to provide an encyclopedic bibliography.

2. This recommendation is distinct from that of Barbot et al. (Citation2019) that researchers using DT measures use a representative range of DT tasks rather than relying only one or two such tasks in their studies, although that recommendation is well-taken.

3. The authors of the 1998 paper cannot remember why they discussed critical issues before recent advances, which makes little temporal or logical sense.

References

  • Acar, S., & Runco, M. A. (2019). Divergent thinking: New methods, recent research, and extended theory. Psychology of Aesthetics, Creativity, and the Arts, 13(2), 153–158. doi:10.1037/aca0000231
  • Aschauer, W., Haim, K., & Weber, C. (2021). A contribution to scientific creativity: A validation study measuring divergent problem solving ability. Creativity Research Journal, 34(2), 195–212.
  • Awofala, A. O., & Fatade, A. O. (2015). Validation of the domains of creativity scale for Nigerian preservice science, technology, and mathematics teachers. Electronic Journal of Research in Educational Psychology, 13(1), 131–150. doi:10.14204/ejrep.35.14057
  • Baer, J. (1993, December/January). Why you shouldn’t trust creativity tests. Educational Leadership, 51(4), 80–83.
  • Baer, J. (1994). Performance assessments of creativity: Do they have long‐term stability? Roeper Review, 17(1), 7–11. doi:10.1080/02783199409553609
  • Baer, J. (1998). The case for domain specificity of creativity. Creativity Research Journal, 11(2), 173–177. doi:10.1207/s15326934crj1102_7
  • Baer, J., & Kaufman, J. C. (2005). Bridging generality and specificity: The amusement park theoretical (APT) model of creativity. Roeper Review, 27(3), 158–163. doi:10.1080/02783190509554310
  • Barbot, B. (2020). Creativity and self‐esteem in adolescence: A study of their domain‐specific, multivariate relationships. The Journal of Creative Behavior, 54(2), 279–292. doi:10.1002/jocb.365
  • Barbot, B., Hass, R. W., & Reiter-Palmon, R. (2019). Creativity assessment in psychological research: (Re)setting the standards. Psychology of Aesthetics, Creativity, and the Arts, 13(2), 233–240. doi:10.1037/aca0000233
  • Basalla, M., Schneider, J., & Vom Brocke, J. (2022). Creativity of deep learning: Conceptualization and assessment. In Proceedings of the 14th International Conference on Agents and Artificial Intelligence. https://www.scitepress.org/Papers/2022/107835/107835.pdf
  • Beaty, R. E., & Johnson, D. R. (2021). Automating creativity assessment with semdis: An open platform for computing semantic distance. Behavior Research Methods, 53(2), 757–780. doi:10.3758/s13428-020-01453-w
  • Becker, M., Cabeza, R., Shaw, H., & Davidson, B. I. (2022). Assessing creativity independently of language: A language-independent remote associate task (LI-RAT). Behavior Research Methods, 54(1), 1–18. doi:10.3758/s13428-021-01585-7
  • Carson, S. H., Peterson, J. B., & Higgins, D. M. (2005). Reliability, validity, and factor structure of the creative achievement questionnaire. Creativity Research Journal, 17(1), 37–50. doi:10.1207/s15326934crj1701_4
  • Ceh, S. M., Edelmann, C., Hofer, G., & Benedek, M. (2022). Assessing raters: What factors predict discernment in novice creativity raters? The Journal of Creative Behavior, 56(1), 41–54. doi:10.1002/jocb.515
  • Cerrato, A., Siano, G., Marco, A. D., & Ricci, C. (2019, June). The importance of spatial abilities in creativity and their assessment through tangible interfaces. In International conference in methodologies and intelligent systems for technology enhanced learning (pp. 89–95). Springer, Cham.
  • Chan, D. W., & Chan, L. K. (1999). Implicit theories of creativity: Teachers’ perception of student characteristics in Hong Kong. Creativity Research Journal, 12(3), 185–195. doi:10.1207/s15326934crj1203_3
  • Clark, P. M., & Mirels, H. L. (1970). Fluency as a pervasive element in the measurement of creativity. Journal of Educational Measurement, 7(2), 83–86. doi:10.1111/j.1745-3984.1970.tb00699.x
  • Clements, D. H. (1991). Enhancement of creativity in computer environments. American Educational Research Journal, 28(1), 173–187. doi:10.3102/00028312028001173
  • Cole, D. G., Sugioka, H. L., & Yamagata-Lynch, L. C. (1999). Supportive classroom environments for creativity in higher education. The Journal of Creative Behavior, 33(4), 277–293. doi:10.1002/j.2162-6057.1999.tb01407.x
  • Couger, J. D., & Dengate, G. (1992, January). Measurement of creativity of IS products. In Proceedings of the Twenty-Fifth Hawaii International Conference on System Sciences (Vol. 4, pp. 288–298). Honolulu, HI, IEEE.
  • Cropley, A. J. (2000). Defining and measuring creativity: Are creativity tests worth using? Roeper Review, 23(2), 72–79. doi:10.1080/02783190009554069
  • Cropley, D. H., & Marrone, R. L. (2021). Automated scoring of figural creativity using a convolutional neural network.
  • Cropley, D. H., Patston, T., Marrone, R. L., & Kaufman, J. C. (2019). Essential, unexceptional and universal: Teacher implicit beliefs of creativity. Thinking Skills and Creativity, 34, 100604. doi:10.1016/j.tsc.2019.100604
  • Cseh, G. M., & Jeffries, K. K. (2019). A scattered CAT: A critical evaluation of the consensual assessment technique for creativity research. Psychology of Aesthetics, Creativity, and the Arts, 13(2), 159. doi:10.1037/aca0000220
  • D’Souza, R. (2021). What characterises creativity in narrative writing, and how do we assess it? Research findings from a systematic literature search. Thinking Skills and Creativity, 42, 100949. doi:10.1016/j.tsc.2021.100949
  • Dawson, P. (2017). Assessment rubrics: Towards clearer and more replicable design, research and practice. Assessment & Evaluation in Higher Education, 42(3), 347–360. doi:10.1080/02602938.2015.1111294
  • Dong, Y., Zhu, S., & Li, W. (2021). Promoting sustainable creativity: An empirical study on the application of mind mapping tools in graphic design education. Sustainability, 13(10), 5373. doi:10.3390/su13105373
  • Dumas, D., Organisciak, P., & Doherty, M. (2021). Measuring divergent thinking originality with human raters and text-mining models: A psychometric comparison of methods. Psychology of Aesthetics, Creativity, and the Arts, 15(4), 645–663. doi:10.1037/aca0000319
  • Eysenck, H. J. (1994). The measurement of creativity. In M. A. Boden (Ed.), Dimensions of creativity (pp. 199–242). The MIT Press.
  • Fardilha, F. D. S., & Allen, J. B. (2020). Defining, assessing, and developing creativity in sport: A systematic narrative review. International Review of Sport and Exercise Psychology, 13(1), 104–127. doi:10.1080/1750984X.2019.1616315
  • Feldhusen, J. F., & Goh, B. E. (1995). Assessing and accessing creativity: An integrative review of theory, research, and development. Creativity Research Journal, 8(3), 231–247. doi:10.1207/s15326934crj0803_3
  • Fluckiger, J. (2010). Single point rubric: A tool for responsible student self-assessment. The Delta Kappa Gamma Bulletin, 76(4), 18. http://www.dkg.org/atf/cf/%7B70E631E4-44B9-4D36-AE7D-D12E4520FAB3%7D/LPB-BulletinSummer2010.pdf
  • Gilbert, F. W., Prenshaw, P. J., & Ivy, T. T. (1996). A preliminary assessment of the effectiveness of creativity training in marketing. Journal of Marketing Education, 18(3), 46–56. doi:10.1177/027347539601800306
  • Glăveanu, V. P. (2015). Creativity as a sociocultural act. The Journal of Creative Behavior, 49(3), 165–180. doi:10.1002/jocb.94
  • Glăveanu, V. P., Hanchett Hanson, M., Baer, J., Barbot, B., Clapp, E. P., Corazza, G. E., & Sternberg, R. J. (2019). Advancing creativity theory and research: A socio‐cultural manifesto. The Journal of Creative Behavior. doi:10.1002/jocb.395
  • Gralewski, J., & Karwowski, M. (2018). Are teachers’ implicit theories of creativity related to the recognition of their students’ creativity? The Journal of Creative Behavior, 52(2), 156–167. doi:10.1002/jocb.140
  • Guilford, J. P. (1967). The nature of human intelligence. McGraw-Hill.
  • Guo, J., Ge, Y., & Pang, W. (2019). The underlying cognitive mechanisms of the rater effect in creativity assessment: The mediating role of perceived semantic distance. Thinking Skills and Creativity, 33, 100572. doi:10.1016/j.tsc.2019.100572
  • Han, J., Forbes, H., & Schaefer, D. (2019, July). An exploration of the relations between functionality, aesthetics and creativity in design. Proceedings of the Design Society: International Conference on Engineering Design, 1(1), 259–268. Cambridge University Press.
  • Han, J., Long, H., Ge, M., & Pang, W. (2022). Perspective-taking feedback: A new feedback affecting creativity. Creativity Research Journal, 34(2), 228–243. doi:10.1080/10400419.2021.1973708
  • Heinen, D. J. P., & Johnson, D. R. (2018). Semantic distance: An automated measure of creativity that is novel and appropriate. Psychology of Aesthetics, Creativity, and the Arts, 12(2), 144–156. doi:10.1037/aca0000125
  • Hester, R. T., & Hester, R. (2012). Scoring collective creativity and legitimizing participatory design. Landscape Journal, 31(1/2), 135–143. doi:10.3368/lj.31.1-2.135
  • Hocevar, D. (1979). Ideational fluency as a confounding factor in the measurement of originality. Journal of Educational Psychology, 71(2), 191–196.
  • Hong, E., Milgram, R. M., & Gorsky, H. (1995). Original thinking as a predictor of creative performance in young children. Roeper Review, 18(2), 147–149. doi:10.1080/02783199509553720
  • Houtz, J. C., & Krug, D. (1995). Assessment of creativity: Resolving a mid-life crisis. Educational Psychology Review, 7(3), 269–300. doi:10.1007/BF02213374
  • Hunsaker, S. L., & Callahan, C. M. (1995). Creativity and giftedness: Published instrument uses and abuses. Gifted Child Quarterly, 39(2), 110–114. doi:10.1177/001698629503900207
  • Kalis, E., & Roke, L. (2011). Adaptation of Runco ideational behavior scale in Latvia. Signum Temporis, 4(1), 36.
  • Kandemir, M. A., & Kaufman, J. C. (2020). The Kaufman domains of creativity scale: Turkish validation and relationship to academic major. The Journal of Creative Behavior, 54(4), 1002–1012. doi:10.1002/jocb.428
  • Kane, M. T. (2016). Explicating validity. Assessment in Education: Principles, Policy & Practice, 23(2), 198–211. doi:10.1080/0969594X.2015.1060192
  • Kapoor, H., Reiter-Palmon, R., & Kaufman, J. C. (2021). Norming the muses: Establishing the psychometric properties of the Kaufman domains of creativity scale. Journal of Psychoeducational Assessment, 39(6), 680–693. doi:10.1177/07342829211008334
  • Kaufman, J. C. (2012). Counting the muses: Development of the Kaufman domains of creativity scale (K-DOCS). Psychology of Aesthetics, Creativity, and the Arts, 6(4), 298. doi:10.1037/a0029751
  • Kaufman, J. C. (2019). Self-assessments of creativity: Not ideal, but better than you think. Psychology of Aesthetics, Creativity, and the Arts, 13(2), 187. doi:10.1037/aca0000217
  • Kaufman, J. C., Plucker, J. A., & Baer, J. (2008). Essentials of creativity assessment. John Wiley & Sons.
  • Kenett, Y. N. (2019). What can quantitative measures of semantic distance tell us about creativity? Current Opinion in Behavioral Sciences, 27, 11–16. doi:10.1016/j.cobeha.2018.08.010
  • Kovalkov, A., Paaßen, B., Segal, A., Pinkwart, N., & Gal, K. (2021). Automatic creativity measurement in scratch programs across modalities. IEEE Transactions on Learning Technologies, 14(6), 740–753. doi:10.1109/TLT.2022.3144442
  • Lehmkuhl, G., Gresse von Wangenheim, C., Martins-Pacheco, L. H., Borgatto, A., & da Cruz Alves, N. (2021). SCORE–A model for the self-assessment of creativity skills in the context of computing education in K-12. Informatics in Education, 20(2), 231–254.
  • Li, Z., Cai, X., Kuznetsova, M., & Kurilova, A. (2022). Assessment of scientific thinking and creativity in an electronic educational environment. International Journal of Science Education, 44(3), 463–486. doi:10.1080/09500693.2022.2032863
  • Lim, W., & Plucker, J. A. (2001). Creativity through a lens of social responsibility: Implicit theories of creativity with Korean samples. The Journal of Creative Behavior, 35(2), 115–130. doi:10.1002/j.2162-6057.2001.tb01225.x
  • Loewenstein, J., & Mueller, J. (2016). Implicit theories of creative ideas: How culture guides creativity assessments. Academy of Management Discoveries, 2(4), 320–348. doi:10.5465/amd.2014.0147
  • Maksić, S. B., & Spasenović, V. Z. (2018). Educational science students’ implicit theories of creativity. Creativity Research Journal, 30(3), 287–294. doi:10.1080/10400419.2018.1488200
  • Malmelin, N., & Nivari-Lindström, L. (2017). Rethinking creativity in journalism: Implicit theories of creativity in the Finnish magazine industry. Journalism, 18(3), 334–349. doi:10.1177/1464884915620272
  • McKay, A. S., Karwowski, M., & Kaufman, J. C. (2017). Measuring the muses: Validating the Kaufman domains of creativity scale (K-DOCS). Psychology of Aesthetics, Creativity, and the Arts, 11(2), 216. doi:10.1037/aca0000074
  • Myszkowski, N., & Storme, M. (2019). Judge response theory? A call to upgrade our psychometrical account of creativity judgments. Psychology of Aesthetics, Creativity, and the Arts, 13(2), 167. doi:10.1037/aca0000225
  • Niu, W., & Sternberg, R. (2002). Contemporary studies on the concept of creativity: The East and the West. The Journal of Creative Behavior, 36(4), 269–288. doi:10.1002/j.2162-6057.2002.tb01069.x
  • O’Loughlin, M. (1992). Rethinking science education: Beyond Piagetian constructivism toward a sociocultural model of teaching and learning. Journal of Research in Science Teaching, 29(8), 791–820. doi:10.1002/tea.3660290805
  • Okuda, S. M., Runco, M. A., & Berger, D. E. (1991). Creativity and the finding and solving or real-world problems. Journal of Psychoeducational Assessment, 9(1), 45–53. doi:10.1177/073428299100900104
  • Orwig, W., Diez, I., Vannini, P., Beaty, R., & Sepulcre, J. (2021). Creative connections: Computational semantic distance captures individual creativity and resting-state functional connectivity. Journal of Cognitive Neuroscience, 33(3), 499–509. doi:10.1162/jocn_a_01658
  • Paek, S. H., Park, H., Runco, M. A., & Choe, H. S. (2016). The contribution of ideational behavior to creative extracurricular activities. Creativity Research Journal, 28(2), 144–148. doi:10.1080/10400419.2016.1162547
  • Paletz, S. B., & Peng, K. (2008). Implicit theories of creativity across cultures: Novelty and appropriateness in two product domains. Journal of Cross-cultural Psychology, 39(3), 286–302. doi:10.1177/0022022108315112
  • Panda, M., & Yadava, R. (2005). Implicit creativity theories in India: An exploration. Psychological Studies, 50(1), 32–39.
  • Patston, T. J., Cropley, D. H., Marrone, R. L., & Kaufman, J. C. (2018). Teacher implicit beliefs of creativity: Is there an arts bias? Teaching and Teacher Education, 75, 366–374. doi:10.1016/j.tate.2018.08.001
  • Pavlović, J., & Maksić, S. (2019). Implicit theories of creativity in higher education: A constructivist study. Journal of Constructivist Psychology, 32(3), 254–273. doi:10.1080/10720537.2018.1477639
  • Plucker, J. A. (1999). Is the proof in the pudding? Reanalyses of Torrance’s (1958 to present) longitudinal data. Creativity Research Journal, 12(2), 103–114. doi:10.1207/s15326934crj1202_3
  • Plucker, J. A., & Alanazi, R. (2019). Is creativity compatible with educational accountability? Promise and pitfalls of using assessment to monitor and enhance a complex construct. In I. Lebuda & V. P. Glǎveanu (Eds.), The Palgrave handbook of social creativity research (pp. 501–514). Cham, Switzerland: Palgrave Macmillan. doi:10.1007/978-3-319-95498-1_31
  • Plucker, J. A., & Beghetto, R. (2004). Why creativity is domain general, why it looks domain specific, and why the distinction does not matter. In R. J. Sternberg, E. L. Grigorenko, & J. L. Singer (Eds.), Creativity: From potential to realization (pp. 153–167). Washington, DC: American Psychological Association.
  • Plucker, J. A., Beghetto, R. A., & Dow, G. T. (2004). Why isn’t creativity more important to educational psychologists? Potentials, pitfalls, and future directions in creativity research. Educational Psychologist, 39(2), 83–96. doi:10.1207/s15326985ep3902_1
  • Plucker, J. A., & Makel, M. C. (2010). Assessment of creativity. In J. C. Kaufman & R. J. Sternberg (Eds.), The Cambridge handbook of creativity (pp. 48–73). Cambridge University Press. doi:10.1017/CBO9780511763205.005
  • Plucker, J. A., McWilliams, J., & Guo, J. (2021). Smart contexts for 21st century talent development. In R. J. Sternberg & D. Ambrose (Eds.), Conceptions of giftedness and talent (pp. 295–316). Palgrave Macmillan.
  • Plucker, J. A., Meyer, M. S., & Makel, M. C. (in press). Supporting innovation: Sociocultural and developmental considerations in the assessment of creativity. In D. Priess, J. C. Kaufman, & M. S. Gonzalez (Eds.), Innovation, creativity, and change across cultures. Palgrave Macmillan.
  • Plucker, J. A., Qian, M., & Schmalensee, S. L. (2014). Is what you see what you really get? Comparison of scoring techniques in the assessment of real-world divergent thinking. Creativity Research Journal, 26(2), 135–143. doi:10.1080/10400419.2014.901023
  • Plucker, J. A., & Renzulli, J. S. (1999). Psychometric approaches to the study of human creativity. In R. J. Sternberg (Ed.), Handbook of creativity (pp. 35–61). Cambridge University Press.
  • Plucker, J. A., & Runco, M. A. (1998). The death of creativity measurement has been greatly exaggerated: Current issues, recent advances, and future directions in creativity assessment. Roeper Review, 21(1), 36–39. doi:10.1080/02783199809553924
  • Plucker, J. A., Runco, M. A., & Lim, W. (2006). Predicting ideational behavior from divergent thinking and discretion- ary time on task. Creativity Research Journal, 18(1), 55–63. doi:10.1207/s15326934crj1801_7
  • Primi, R., Silvia, P. J., Jauk, E., & Benedek, M. (2019). Applying many-facet Rasch modeling in the assessment of creativity. Psychology of Aesthetics, Creativity, and the Arts, 13(2), 176–186. doi:10.1037/aca0000230
  • Puccio, G. J., & Chimento, M. D. (2001). Implicit theories of creativity: Laypersons’ perceptions of the creativity of adaptors and innovators. Perceptual and Motor Skills, 92(3), 675–681. doi:10.2466/pms.2001.92.3.675
  • Qian, M., & Plucker, J. A. (2018). Looking for renaissance people: Examining domain specificity-generality of creativity using item response theory models. Creativity Research Journal, 30(3), 241–248. doi:10.1080/10400419.2018.1488348
  • Qian, M., Plucker, J. A., & Yang, X. (2019). Is creativity domain specific or domain general? Evidence from multilevel explanatory item response theory models. Thinking Skills and Creativity, 33, 100571. doi:10.1016/j.tsc.2019.100571
  • Reiter-Palmon, R., Forthmann, B., & Barbot, B. (2019). Scoring divergent thinking tests: A review and systematic framework. Psychology of Aesthetics, Creativity, and the Arts, 13(2), 144–152. doi:10.1037/aca0000227
  • Romero, M., Arnab, S., De Smet, C., Mohamad, F., Minoi, J. L., & Morini, L. (2019). Assessment of co-creativity in the process of game design. Electronic Journal of e-Learning, 17(3), 199–206. doi:10.34190/JEL.17.3.003
  • Runco, M. A. (2007). To understand is to create: An epistemological perspective on human nature and personal creativity. In R. Richards (Ed.), Everyday creativity and new views of human nature: Psychological, social, and spiritual perspectives (pp. 91–107). American Psychological Association. doi:10.1037/11595-004
  • Runco, M. A., & Albert, R. S. (1985). The reliability and validity of ideational originality in the divergent thinking of academically gifted and nongifted children. Educational and Psychological Measurement, 45(3), 483–501.
  • Runco, M. A., & Jaeger, G. J. (2012). The standard definition of creativity. Creativity Research Journal, 24(1), 92–96. doi:10.1080/10400419.2012.650092
  • Runco, M. A., Millar, G., Acar, S., & Cramond, B. (2010). Torrance tests of creative thinking as predictors of personal and public achievement: A fifty-year follow-up. Creativity Research Journal, 22(4), 361–368. doi:10.1080/10400419.2010.523393
  • Runco, M. A., & Mraz, W. (1992). Scoring divergent thinking tests using total ideational output and a creativity index. Educational and Psychological Measurement, 52(1), 213–221. doi:10.1177/001316449205200126
  • Runco, M. A., Noble, E. P., Reiter-Palmon, R., Acar, S., Ritchie, T., & Yurkovich, J. M. (2011). The genetic basis of creativity and ideational fluency. Creativity Research Journal, 23(4), 376–380. doi:10.1080/10400419.2011.621859
  • Runco, M. A., Okuda, S. M., & Thurston, B. J. (1987). The psychometric properties of four systems for scoring divergent thinking tests. Journal of Psychoeducational Assessment, 5(2), 149–156. doi:10.1177/073428298700500206
  • Runco, M. A., Plucker, J. A., & Lim, W. (2001). Development and psychometric integrity of a measure of ideational behavior. Creativity Research Journal, 13(3–4), 393–400.
  • Runco, M. A., Walczyk, J. J., Acar, S., Cowger, E. L., Simundson, M., & Tripp, S. (2014). The incremental validity of a short form of the ideational behavior scale and usefulness of distractor, contraindicative, and lie scales. The Journal of Creative Behavior, 48(3), 185–197. doi:10.1002/jocb.47
  • Said-Metwaly, S., Fernández-Castilla, B., Kyndt, E., Van den Noortgate, W., & Barbot, B. (2021). Does the fourth-grade slump in creativity actually exist? A meta-analysis of the development of divergent thinking in school-age children and adolescents. Educational Psychology Review, 33(1), 275–298. doi:10.1007/s10648-020-09547-9
  • Sawyers, J. K., & Canestaro, N. C. (1989). Creativity and achievement in design coursework. Creativity Research Journal, 2(1–2), 126–133. doi:10.1080/10400418909534306
  • Sen, S. (2016). Applying the mixed Rasch model to the Runco ideational behavior scale. Creativity Research Journal, 28(4), 426–434. doi:10.1080/10400419.2016.1229985
  • Sen, S. (2022). A reliability generalization meta-analysis of Runco ideational behavior scale. Creativity Research Journal, 34(2), 178–194. doi:10.1080/10400419.2021.1960719
  • Shute, V. J., & Rahimi, S. (2021). Stealth assessment of creativity in a physics video game. Computers in Human Behavior, 116, 106647. doi:10.1016/j.chb.2020.106647
  • Silvia, P. J. (2011). Subjective scoring of divergent thinking: Examining the reliability of unusual uses, instances, and consequences tasks. Thinking Skills and Creativity, 6(1), 24–30. doi:10.1016/j.tsc.2010.06.001
  • Silvia, P. J. (2018). Creativity is undefinable, controllable, and everywhere. In R. J. Sternberg & J. C. Kaufman (Eds.), The nature of human creativity (pp. 291–301). Cambridge University Press.
  • Silvia, P. J., Martin, C., & Nusbaum, E. C. (2009). A snapshot of creativity: Evaluating a quick and simple method for assessing divergent thinking. Thinking Skills and Creativity, 4(2), 79–85. doi:10.1016/j.tsc.2009.06.005
  • Silvia, P. J., Rodriguez, R. M., Beaty, R. E., Frith, E., Kaufman, J. C., Loprinzi, P., & Reiter-Palmon, R. (2021). Measuring everyday creativity: A Rasch model analysis of the Biographical Inventory of Creative Behaviors (BICB) scale. Thinking Skills and Creativity, 39, 100797. doi:10.1016/j.tsc.2021.100797
  • Silvia, P. J., Wigert, B., Reiter-Palmon, R., & Kaufman, J. C. (2012). Assessing creativity with self-report scales: A review and empirical evaluation. Psychology of Aesthetics, Creativity, and the Arts, 6(1), 19–34.
  • Silvia, P. J., Winterstein, B. P., Willse, J. T., Barona, C. M., Cram, J. T., Hess, K. I., & Richard, C. A. (2008). Assessing creativity with divergent thinking tasks: Exploring the reliability and validity of new subjective scoring methods. Psychology of Aesthetics, Creativity, and the Arts, 2(2), 68–85. doi:10.1037/1931-3896.2.2.68
  • Simonton, D. K. (2012). Taking the U.S. Patent Office criteria seriously: A quantitative three-criterion creativity definition and its implications. Creativity Research Journal, 24(2–3), 97–106. doi:10.1080/10400419.2012.676974
  • Simonton, D.K. 2019. The sociocultural context of exceptional creativity: Historiometric studies. In I. Lebuda & V. Glăveanu (Eds.), The Palgrave handbook of social creativity research. Palgrave Macmillan, Cham. https://doi.org/10.1007/978-3-319-95498-1_12
  • Snyder, H. T., Hammond, J. A., Grohman, M. G., & Katz-Buonincontro, J. (2019). Creativity measurement in undergraduate students from 1984–2013: A systematic review. Psychology of Aesthetics, Creativity, and the Arts, 13(2), 133–143. doi:10.1037/aca0000228
  • Spiel, C., & von Korff, C. (1998). Implicit theories of creativity: The conceptions of politicians, scientists, artists and school teachers. High Ability Studies, 9(1), 43–58. doi:10.1080/1359813980090104
  • Stefanic, N., & Randles, C. (2015). Examining the reliability of scores from the consensual assessment technique in the measurement of individual and small group creativity. Music Education Research, 17(3), 278–295. doi:10.1080/14613808.2014.909398
  • Sternberg, R. J. (1985). Implicit theories of intelligence, creativity, and wisdom. Journal of Personality and Social Psychology, 49(3), 607–627. doi:10.1037/0022-3514.49.3.607
  • Sung, Y. T., Cheng, H. H., Tseng, H. C., Chang, K. E., & Lin, S. Y. (2022). Construction and validation of a computerized creativity assessment tool with automated scoring based on deep-learning techniques. Psychology of Aesthetics, Creativity, and the Arts. doi:10.1037/aca0000450
  • Susanto, E., Novitasari, Y., Rakhmat, C., Hidayat, A., & Wibowo, S. B. (2018, November). Rash model analysis of Kaufman domains of creativity scale (K-DOCS) to Indonesian students. Journal of Physics: Conference Series, 1114(1), 12027. IOP Publishing.
  • Tan, M., Mourgues, C., Bolden, D. S., & Grigorenko, E. L. (2014). Making numbers come to life: Two scoring methods for creativity in Aurora’s cartoon numbers. The Journal of Creative Behavior, 48(1), 25–43. doi:10.1002/jocb.39
  • Tavani, J. L., Caroff, X., Storme, M., & Collange, J. (2016). Familiarity and liking for music: The moderating effect of creative potential and what predict the market value. Learning and Individual Differences, 52, 197–203. doi:10.1016/j.lindif.2014.11.026
  • Tep, P., Maneewan, S., & Chuathong, S. (2021). Psychometric Examination of Runco Ideational Behavior Scale: Thai Adaptation. Psicologia: Reflexão e Crítica, 34, 1-11.
  • Tsai, K. C. (2015). Assessing a Chinese version of the Runco ideational behavior scale. Social Behavior and Personality: An International Journal, 43(7), 1111–1122. doi:10.2224/sbp.2015.43.7.1111
  • van Broekhoven, K., Cropley, D., & Seegers, P. (2020). Differences in creativity across art and STEM students: We are more alike than unalike. Thinking Skills and Creativity, 38, 100707. doi:10.1016/j.tsc.2020.100707
  • Van Hooijdonk, M., Mainhard, T., Kroesbergen, E. H., & Van Tartwijk, J. (2022). Examining the assessment of creativity with generalizability theory: An analysis of creative problem solving assessment tasks. Thinking Skills and Creativity, 43, 100994. doi:10.1016/j.tsc.2021.100994
  • Vartanian, O. (2014). Toward a cumulative psychological science of aesthetics, creativity, and the arts. Psychology of Aesthetics, Creativity, and the Arts, 8(1), 15–17. doi:10.1037/a0035723
  • Wallach, M. A. (1976, January-February). Tests tell us little about talent. American Scientist, 64(1):57–63.
  • Wang, J., & Long, H. (2022). Reexamining subjective creativity assessments in science tasks: An application of the rater-mediated assessment framework and many-facet Rasch model. Psychology of Aesthetics, Creativity, and the Arts. doi:10.1037/aca0000470