2,859
Views
3
CrossRef citations to date
0
Altmetric
Articles

A Summary of Construct Validity Evidence for Two Measures of Character Strengths

ORCID Icon
Pages 302-313 | Received 23 Jan 2022, Accepted 21 Aug 2022, Published online: 19 Sep 2022

Abstract

The VIA Inventory of Strengths has become the most widely used instrument in the world for measuring the construct referred to character strengths. However, several limitations were noted in its original development. In response, the VIA Assessment Suite for Adults was developed as a battery of instruments intended to address those gaps. The suite includes two inventories providing dimensional measures of the character strengths: the VIA Inventory of Strengths-Revised and the Global Assessment of Character Strengths. Short forms were also developed for each. So far, five reasonably sized samples of adults (total N = 7,924) have provided evidence for the empirical validity of some subset of these instruments, making them the most thoroughly vetted measures of character strengths available today. This article aggregates previously available and new findings on their construct validity. Evidence concerning substantive validity, structural validity, and external validity is reviewed, and in some cases aggregated across samples. The findings generally support the construct validity of the instruments evaluated according to all three standards, with exceptions noted. Recommendations are offered for their use in research and applied settings.

The VIA Classification of Character Strengths and Virtues (Peterson & Seligman, Citation2004) is a structural model that conceptualizes positive human traits in terms of 24 character strengths, as well as six superordinate variables developed conceptually that are referred to as the virtues (see ). The model’s development was one of the seminal events in the founding of positive psychology. The success of the VIA model can be attributed to several factors beyond its primacy, though. First, its development involved input from more than 50 experts in positive human functioning as well as reviews of a variety of literatures, providing a broad foundation for the dimensions chosen for inclusion. This strategy was in contrast to earlier models of positive functioning that were generally the work of one person or at most several people. Second, the model introduced a concept referred to as signature strengths. These are defined as a subset of the strengths that the individual identifies as particularly central to his or her identity, the exercise of which is experienced as particularly energizing or fulfilling by the individual. The concept of signature strengths has played a central role in the application of the model to coaching, clinical, educational, and corporate settings (e.g., Linley et al., Citation2010; Merritt et al., Citation2019; Suldo et al., Citation2015; Uliaszek et al., Citation2022). Finally, these and other studies have connected character and signature strengths to flourishing and well-being in a variety of contexts (e.g., Blanchard et al., Citation2020).

Table 1. The VIA classification of character strengths and virtues.

One of the most important contributors to the popularity of the model was the early introduction of inventories for measuring the 24 VIA character strengths. The first to be developed was the VIA Inventory of Strengths (VIA-IS; Peterson & Seligman, Citation2004), intended for use with adults 18 and over. This was followed shortly thereafter by the VIA-Youth (Park & Peterson, Citation2006), for youth ages 10–17. These inventories treat the 24 character strengths as personality traits (McGrath et al., Citation2020) capable of being measured as relatively enduring though potentially malleable attributes of the individual. The VIA-IS and VIA-Youth have also been used extensively to identify signature strengths in applied settings based on which scores are highest. The scale scores provide a basis for nomothetic evaluation of the results, while the signature strengths can be used as the basis for a more idiographic approach to interpretation. The VIA-IS and VIA-Youth have since been completed millions of times online and have been used in hundreds of studies.

In 2014, the VIA Institute on Character, the copyright holder for the two instruments, decided to fund revision of the VIA-IS. The reasons for this decision have been summarized elsewhere (McGrath, Citation2019), so the primary considerations will be listed here only in brief:

  • All the items were positively keyed, which is considered undesirable by many testing professionals because of the potential for yea-saying (“all true”) or nay-saying (“all false”) response sets (McGrath et al., Citation2010).

  • At 240 items (10 items per strength scale), the instrument was also too long for practical purposes in many contexts. For several years, a 120-item version of the VIA-IS was administered at the VIA Institute website while the VIA-IS was under revision. This version, called the VIA-120, consisted of the five items out of the original ten on each scale with the highest corrected item-total correlations in a sample of N = 458,854 who completed the VIA-IS online. However, users of the test were requesting even briefer versions.

  • The original VIA-IS did not include measures of the six virtues. Some researchers have generated proxy virtue scores by aggregating scores across the scales comprising the virtues (e.g., Gustems-Carnicer & Calderón, Citation2016), but this strategy was never encouraged by the developers of the model and has never been empirically evaluated. Subsequent research on the latent structure of the 24 strengths has suggested a three-factor model that appears reliably across populations and measurement instruments (McGrath, Citation2015; McGrath et al., Citation2018), and that are consistent with the concept of higher-order virtues (McGrath, Citation2021). The primary markers for these latent variables are provided in . The popularity of the original virtue model and the empirical support for the revised model suggested the potential value of scales specifically targeting both virtue models.

  • Problems were identified with the item content of several scales. These included questions asking about protected health information or other sensitive topics, and scales that included items reflecting markedly heterogeneous contents.

The VIA Assessment Suite for Adults

In response to these concerns, an extensive revision process was pursued that continued until 2019. This process ultimately resulted in the development of three measurement tools, collectively referred to as the VIA Assessment Suite for Adults (McGrath, Citation2019).

The premier instrument in this new set is the VIA-Inventory of Strengths-Revised (VIA-IS-R). Major differences between the VIA-IS and the VIA-IS-R include the following:

  • The latter instrument consists of 192 items (eight items per strength), compared to 240 on the VIA-IS.

  • The VIA-IS-R includes both positively and negatively keyed items on each scale.

  • Items were selected based on a number of criteria. These included reliability statistics, results from item response theory analyses, ratings of item prototypicality for the scale construct (Broughton, Citation1984), and item readability statistics.

  • Virtue scales were developed representing both the original six-virtue model and the evidence-based three-virtue model. Their addition raised the number of scales that can be scored from 24 to 33. The virtue scales can also be administered independently of the VIA-IS-R.

  • Two short forms of the VIA-IS-R were developed simultaneously, both of which are 96 items long (four items per strength), which is consistent with the typical number of items on measures used in applied research (Flake et al., Citation2017). The VIA-IS-M (Mixed) version includes both positively and negatively keyed items like the VIA-IS-R. The VIA-IS-P (Positive) consists of items that are all positively keyed, reducing the cognitive load for the respondent. Virtue scales are also available for the short forms. The three versions as a group will be referred to here as the VIA-IS-R/M/P.

Items on the VIA-IS-R/M/P are completed on a 1 (Very Much Unlike Me) to 5 (Very Much Like Me) scale. Items are averaged so that all scale scores are on a 1–5 scale. Researchers and practitioners interested in collecting data using the VIA-IS-R/M/P are asked to request permission from the VIA Institute (https://www.viacharacter.org/www/Research/Conduct-a-Study). This requirement was implemented primarily to track the use of the instrument. No reasonable request is denied, and there is no cost for the inventory’s use.

In addition to the VIA-IS-R/M/P, the Assessment Suite includes a new inventory providing dimensional measures of the 24 strengths. The Global Assessment of Character Strengths (GACS-72; McGrath, Citation2019) is a 72-item questionnaire, three items per strength. Respondents first see a brief description of each of the 24 strengths. They then complete 24 items asking whether each strength is “essential” to them on a 1 (Very Strongly Disagree) to 7 (Very Strongly Agree) scale. These are followed by 24 items asking whether the strength is “natural and effortless” for them to use, and another 24 asking whether the strength is “energizing.” This wording comes from Peterson and Seligman’s (Citation2004) discussion of how individuals they interviewed characterized their signature strengths. The 24 “natural and effortless” items can also be used as single-item indicators of each strength, bundled as the GACS-24. The two versions in combination can be referred to as the GACS-72/24.

The GACS-72 may be found in McGrath (Citation2019) as well as on the VIA Institute website. It is freely available for researchers and practitioners to modify for their personal purposes and use without approval from the VIA Institute. The Institute will also provide data collection online for any of their measures in their original form at no charge.Footnote1

The validation process

Flake et al. (Citation2017) have suggested construct validation of scale scores consists of three elements. The first involves developing the instrument with issues of substantive validity in mind. The substantive validity of an instrument depends upon two factors. The first has to do with the degree to which the constructs to be measured are clearly defined. The second has to do with the degree to which items are chosen that effectively reflect those constructs. These are primarily conceptual issues.

The second and third elements of validation in this model are primarily empirical in focus. Structural validation has to do with the psychometric adequacy of instrument scores. Hussey and Hughes (Citation2020) suggested four key indicators for structural validity: internal reliability, test-retest reliability, confirmatory factor structure, and measurement invariance. External validation is the term Flake et al. (Citation2017) used to refer to the degree to which scale scores are related to other variables in ways that match expectations. They described several approaches to the evaluation of external validity, all of which represented aspects of convergent and discriminant validity.

Flake et al. (Citation2017) and Hussey and Hughes (Citation2020) raised concerns about the extent to which many studies rely on measures for which there is insufficient evidence of validity. In many cases an estimate of internal reliability is the only justification provided for using an instrument. Even this minimal evaluation is usually accomplished using coefficient alpha, a statistic that is considered questionable because of its unrealistic assumptions about the component items (e.g., McNeish, Citation2018; Sijtsma, Citation2009). These concerns are relevant to the current status of measurement in positive psychology research. Even articles introducing new measures of constructs in positive psychology often offer little more statistical support for their use than internal reliability metrics. For many measures, evidence of external validity is often missing all together.

Turning to the construct validity of the VIA Assessment Suite, several factors argue for the substantive validity of the new instruments. The original choice and definition of the 24 strengths was based on multiple literature reviews and substantial expert input. Extensive descriptions and nomological models are available for each of the 24 character strengths (Peterson & Seligman, Citation2004), the original six virtues (Dahlsgaard et al., Citation2005), and the three empirically derived virtues (McGrath et al., Citation2018). These sources were extensively used in the development of new items. Explicit definitions of the character strengths are even incorporated into the GACS-72/24. Item prototypicality ratings, item unidimensionality, and item-total correlations were all considered in the selection of items for the VIA-IS-R/M/P, which should contribute to their substantive validity as well.

That said, concerns can still be raised about the degree to which the items of the VIA-IS-R/P/M and GACS-72/24 provide adequate coverage for all the character strengths. Arbenz et al. (in press) asked psychologists to rate the breadth of the character strengths on a 1–9 scale. In every case the mean exceeded 5.0 (M = 6.13, range = [5.13, 7.09], raising questions about whether even eight-item scales would be sufficient for adequate coverage across all strengths. This concern reflects the practical obstacles to balancing competing ideals in psychological measurement. It would be desirable for scales to include items tapping all elements of a content domain. They should also be brief to avoid taxing respondents, particularly when the number of constructs being measured is large as in the case of the 24 character strengths. At the same time, it has been argued by some that the most important feature of multi-item scales is unidimensionality (e.g., Ng et al., Citation2017). What can be said about the substantive validity of the two instruments is that the items are likely to be representative of the underlying constructs but may not necessarily be comprehensive.

Moving on from substantive validity, data are now available from five samples of adults who completed the VIA-IS-R/M/P, four of which also took the GACS-72/24. These data can be used to evaluate the empirical validity of these instruments. The remainder of this manuscript will use results from these five samples to generate a comprehensive summary of the two instruments’ structural and external validity. The next section provides a description of the five currently available validation samples.

Description of the samples

The five samples collectively include 7,924 adults ages 18 and older. All data collection was approved by the Fairleigh Dickinson University Institutional Review Board or was judged exempt from review. Attention checks were implemented in all samples but the first.

The first sample will be called the VIA Derivation Sample (VD Sample; McGrath, Citation2019). Between October 2015 and March 2016, individuals who accessed the VIA Institute website and completed the English language version of the VIA-120 in return for personal feedback on their results were asked to volunteer to complete additional questionnaires for research purposes. This request generated a sample of 4,286 adults who responded to the entire set of materials. The sample was 77.7% female, with an average age of 45.6 years (SD = 13.1). They represented 83 different countries, though the majority were from the United States (50.9%), Australia (10.9%), Canada (7.4%), and the United Kingdom (6.0%). In addition to demographic variables, participants completed an additional 309 candidate items for the VIA-IS-R as well as the GACS-72, the SSS, and a 48-item questionnaire consisting of two behavioral acts representing each of the 24 strengths. This sample was used to select the items for the VIA-IS-R and its short forms.

The VIA Cross-Validation Sample (VC Sample; McGrath & Wallace, Citation2021) consisted of 631 U.S. residents who also completed the English version of the VIA-120 at the VIA Institute website, this time between August and October 2017, and subsequently agreed to participate in a study. Participants were 76.9% female, with a mean age of 41.9 years (SD = 13.1). They were administered the three VIA Assessment Suite instruments in their entirety (they were warned there would be some duplication of items with the VIA-120) as well as a 168-item questionnaire about the frequency of seven behavioral acts conceptually representing each of the 24 strengths. For the VC and subsequent sample, 17 attention items were added using the general format “select option x for this item.” Individuals who failed more than three attention items were excluded from the final analysis.

The Mechanical Turk Cross-Validation Sample (MC Sample; McGrath & Wallace, Citation2021) included 743 Amazon Mechanical Turk workers from the U.S. who completed the VIA Assessment Suite measures as well as the same behavioral acts measure completed by the VC Sample. Initial data collection occurred in June 2017. This sample was 49.7% female, with an average age of 34.4 (SD = 10.2). MC Sample members who completed all the measures and passed the attention check as described for the VC Sample were contacted three months later and invited to complete the VIA-IS-R again for purposes of evaluating test-retest reliability. The Time 2 sample included 474 workers. Bonferroni-corrected chi square tests revealed no significant differences between those who did and did not respond to the Time 2 invitation on any categorical demographic variable, with the same result for Bonferroni-corrected t tests of age and the 33 VIA-IS-R scores at Time 1. Members of the MC Sample received compensation for their participation even if they failed the attention check. The publication summarizing results for the VC and MC samples focused on the VIA-IS-R and its short forms, but statistics for the GACS in these samples will also be provided in the present article.

Fourth, between October and November 2019, a stratified sample of 1,765 U.S. residents was gathered in collaboration with Qualtrics, with the goal of matching U.S. adult Census statistics for gender, age, education, self-identified racial identity, and region of the country (McGrath et al., Citation2022). This Qualtrics Representative Sample (QR Sample) completed the VIA Assessment Suite instruments as well as several other questionnaires used in another study. The sample was 51.8% female, with an average age of 46.5 (SD = 17.0). Behavioral criteria specific to the character strengths were not included in this study. In consultation with Qualtrics, the number of attention items was reduced to five, with participants excluded if they failed more than two. Participants were compensated, with the amount varying depending on the difficulty associated with recruiting members of certain demographic subgroups.

Finally, a sample that will be called the German Cross-Validation Sample (GC Sample; Vylobkova et al., in press) consisted of 499 adults recruited via mailing lists and social media to complete the German-language versions of the VIA-IS-R and the original VIA-IS. The sample was 79% female, with an average age of 33.3 years (SD = 13.9). Where the previously listed studies involved completing the entire VIA-IS-R, the authors of this study used a planned missing data design. The protocol involved discrete administration of the VIA-IS and VIA-IS-R. However, full administration of both inventories was deemed too long (240 items for the VIA-IS and 192 for the VIA-IS-R). Instead, each participant was administered randomly chosen blocks of items from the two inventories in a sequence that ensured they were administered half the items from each of the 24 strength scales from each inventory. Unfortunately, the method of data collection does not allow for identifying which items were administered to which participants (V. Vylobkova, personal communication, June 7, 2022), so new short forms and virtue scales could not be computed, and additional item analyses could not be conducted.

The order of administration of the two inventories was also counterbalanced. Participants completed several other measures, one of which was used in this review. The Core Virtue Rating Form (Ruch et al., Citation2020) consists of six items, each of which represents one of the six virtues that were part of the original VIA Classification (see ). The items begin with a short description of the virtue based on Peterson and Seligman (Citation2004). The respondent then indicates how well the description fits their typical behaviors on a scale from 1 (not at all) to 9 (absolutely).

In return for participating, students received course credit. All participants received feedback on their results, and a small donation was made to a charitable organization for each completion of the study. Several variables (e.g., speed of completion, number of missing items) were used to exclude participants from the analyses.

Evidence for reliability

For each of the multi-item scales that comprise the VIA-IS-R/M/P and the GACS-72, two estimates of internal reliability were generated. Coefficient alpha was computed to provide a comparison with results for other instruments. Given the criticisms that have been leveled against that statistic, McDonald’s (Citation1999) omega total was also computed.

Since the GACS-24 consists of single-item scales, an alternative strategy was needed to estimate reliability. Wanous and Hudy (Citation2001) suggested such a method based on the correction for attenuation: rxx=rxy2ryy.

In this formula, rxx is the reliability of a single-item measure, ryy is the reliability of a standard measure of the same latent construct, and rxy is the correlation between the two measures. In this case, ryy was estimated by computing coefficient alpha using the “essential” and “energizing” items for each strength. Alpha was used because omega could not be computed with only two items, and the concerns about alpha are reduced somewhat given the very high similarity in the two items used in the computation.

Internal reliability estimates for the first four samples were computed using the MBESS (Kelley, Citation2020) package in R (R Core Team, Citation2021), while reliability statistics from the article on the GC sample were computed with the psych package (Revelle, Citation2020). provides mean reliability values across the samples weighted by sample size. Results were similar for the two reliability statistics. Specifically, 76% of mean reliability estimates for the VIA-IS-R were ≥ .80, while 100% were ≥ .70. This distribution is consistent with a large-scale review of reliability values across studies (Peterson et al., Citation2020). All reliability estimates for the GACS-72 also were .80 or higher.

Table 2. Mean internal reliability values across samples.

Unsurprisingly, results were weaker for the shorter measures. Only 21% of omega coefficients for the VIA-IS-M, 39% for the VIA-IS-P, and 21% for the GACS-24 were ≥ .80. These numbers increased to 85%, 88%, and 96%, respectively, when .70 was used as a minimum value, which Cortina (Citation1993) suggested as the acceptable minimum for reliability. Finally, reliability estimates for the VIA-IS-M Judgment scale was slightly less than .60, which is the lowest recommended value for a reliability statistic according to other sources (e.g., Shrout, Citation1998).

The second pillar of structural validity (Hussey & Hughes, Citation2020) is test-retest reliability. As noted previously, members of the MC Sample provided three-month retest data for the VIA-IS-R. Test-retest reliability was estimated for each of the 33 scales of the VIA-IS-R/M/P inventories using the mixed model absolute agreement intraclass correlation for a single measure. McGrath and Wallace (Citation2021, Table 8) provided all 99 estimates, so results will only be summarized here. Reliability was again strongest for the VIA-IS-R, with 76% at .80 or higher, and all were .70 or higher. In contrast, only 21% of VIA-IS-M correlations and 36% of VIA-IS-P correlations were ≥ .80. All equaled .70 or greater except in three cases: the VIA-IS-M Humor scale (.68), and the VIA-IS-P Humility (.65) and Judgment (.67) scales.

A third approach to exploring reliability in the context of the VIA-IS-R/M/P instruments is made possible by the 1–5 scaling of all the subtests resulting from the use of item averages for scores. McGrath et al. (Citation2022) took advantage of this commonality to evaluate the alternate versions reliability of each of the VIA-IS-R/M/P scales across the three versions of the inventory in the QR Sample. For the present study, this analysis was extended to all four samples, and to the GACS-72/24 inventories as well. Mixed model absolute agreement intraclass correlations for a single measure for each scale may be found in . In this case, 91% of mean reliability estimates across the VIA-IS-R/M/P inventories were .80 or higher, and all were ≥ .70. All estimates across the GACS-72/24 scales exceeded .80. These findings suggest the three versions can be treated as essentially interchangeable instruments. This conclusion should be accompanied by the caution that the short forms were computed by extracting the short form items from the full item set, which would potentially exaggerate their convergence (see Smith et al., Citation2000).

Table 3. Summary of alternate versions reliability within samples.

Latent structural analyses

Tests of unidimensionality

In addition to internal reliability and test-retest reliability, Hussey and Hughes (Citation2020) associated structural validity with a unidimensional factor structure and measurement invariance. Although some have argued that unidimensionality is an important if not essential feature for a well-functioning psychometric instrument (e.g., Feraco et al., Citation2022; Ng et al., Citation2017), others have argued this may be an unrealistic standard for useful measures of personality (Hopwood & Donnellan, Citation2010). This argument is particularly relevant to the VIA Classification constructs. Peterson and Seligman (Citation2004) conceptualized some of the strengths as inherently multidimensional. As indicated by the text in brackets in , many of the strengths were given alternate labels to reflect this complexity. The virtue scales were also explicitly developed in a manner that included items from each of the strengths most closely related to the virtue, so unidimensionality was never intended as a consideration in their development (McGrath, Citation2019).

With these caveats in mind, one-factor confirmatory factor analyses were generated for the 24 strength scales in the four samples. These were conducted using the WLSMV estimator in the R lavaan package (Rosseel, Citation2012). Hu and Bentler (Citation1999) evaluated a variety of rules for establishing good fit. Their results indicated using a squared root mean residual (SRMR) > .09 combined with a comparative fit index (CFI) < .95 as an indicator of poor fit. Out of 384 analyses (24 strengths from four inventories in four samples), 34 did not initially meet both these criteria, none of which involved the GACS-72. To address this issue, one error covariance was freed based on the largest modification index. This resolved the issue for 23 models, including all instances involving the VIA-IS-M and VIA-IS-P. The process was repeated, reducing the number of scales for which good fit was not achieved to two, the VIA-IS-R Bravery and Humility scales in the QR sample. In these cases, freeing a third error covariance resolved the issue. Overall, 47 out of 4,128 error covariances had to be freed to achieve good fit in all the models. Interestingly, 18 of the 39 freed covariances in the VC, MC, and RS samples involved consecutively administered items on a scale, suggesting item effects may have contributed to covariation not attributable to the shared factor even though items from each scale were administered at 24-item intervals to those samples. provides the range of final estimates for the two statistics across the four samples.

Table 4. Tests of scale unidimensionality across samples.

Measurement invariance

The evaluation of measurement invariance across subgroups is treated as a sequential process. The first step examines the configural invariance of a factor structure across subgroups. Evidence of invariance suggests the model is equally applicable across the populations represented by the groups. Step 2 tests metric invariance by fixing corresponding loadings to be equal across groups. In this case, invariance provides a basis for allowing comparisons of covariation in the subpopulations. Finally, scalar invariance is evaluated by fixing intercepts for corresponding variables to be equal across groups. Scalar invariance provides evidence that mean differences in the two subpopulations represent differences in the underlying factor(s) rather than resulting from extraneous causes.

In the present study, invariance was evaluated applying the unidimensional models evaluated in the previous section across genders and age groups. Cases with missing data for the demographic variable were omitted, as were respondents who identified as transgender from the gender analyses. The age analyses were based on a median split of participants within each sample into younger and older subgroups. For purposes of consistency with the previous analyses, error covariances needed to achieve good fit for the unidimensional model were allowed to vary. In addition, models were evaluated for good fit using the same criteria used for unidimensionality (SRMR ≤ .09, CFI ≥ 0.95).

summarizes the results for both gender and age. For each strength scale, four values are provided. The first is the number of samples out of four in which the test of configural invariance failed, that is, basic factor structure seemed to vary across the two groups. The second number is the number of samples in which only configural invariance was demonstrated, that is, the test of configural invariance was passed but not the test of metric invariance. The third number is instances where metric invariance was demonstrated based on equivalent factor loadings across groups. The final number indicates how frequently scaler invariance was achieved. Again, scalar and metric invariance required good fit in prior models as well.

Table 5. Tests of measurement invariance across samples.

Out of 768 tests of configural invariance, only one demonstrated poor fit (the VIA-IS-R Curiosity scale for age in the MC Sample). Because scalar invariance is required for comparisons of means across populations to be clearly interpretable, it is considered particularly important. For both gender and age, four VIA-IS-R strengths failed to achieve scalar invariance in at least three samples: Bravery, Curiosity, Humility, and Judgment for gender; and Bravery, Curiosity, Honesty, and Humility for age. None of the scales of the VIA-IS-P or GACS-72 failed to meet this standard. In fact, scalar invariance was demonstrated for every GACS-72 scale in every sample, and in all but four VIA-IS-P analyses. The pattern demonstrates a connection between number of items per scale, reverse keying, and invariance outcomes. Even so, the results were strongly supportive of the interpretability of mean differences on all but a few scales. In addition, at least metric invariance was demonstrated in the majority of samples for every scale, suggesting relationships with other variables can be interpreted across ages and genders.

Evidence for external validity

As noted previously, three of the samples (the VD, VC, and MC Samples) also completed a behavioral measure of the character strengths and virtues. These criteria were developed using the act frequency approach to measuring personality variables (Buss & Craik, Citation1983) by consensus among the researchers involved in each study. The GC Sample also completed an item for each of the six virtues in the original VIA Classification. These measures can be used to gauge the convergent and discriminant validity of the Assessment Suite instruments, but certain caveats should be raised about doing so. First, the term “convergent validity” is being used broadly here, with no implication that the criteria used in the earlier studies converge with the strength scales on a common latent variable. The virtue scales used in the GC Sample are at best seen as potential correlates of the strength scales. The behavioral scales used in the other samples similarly should not be assumed to be convergent measures in the absence of any evidence for their validity. This point raises a second concern, which is that none of these “criterion” measures have themselves been externally validated. Third, both the strength scales and the act scales tend to demonstrate substantial intercorrelations, limiting the potential for achieving discriminant validity. The virtue measure in the GC Sample is also problematic as a criterion in that the original association between strength and virtue was developed conceptually by Peterson and Seligman (Citation2004). Several studies have now raised questions about some of these assignments (see Ruch et al., Citation2020, for a review). In particular, Humor seems to have more to do with the virtue of Humanity, Hope may fit better as an exemplar of Courage, and Leadership and Honesty each seems interstitial between virtues, as will be discussed below.

With these caveats in mind, summarizes the results for the external validity analyses. For each instrument, the convergent validity column provides the mean correlation weighted by sample size between the strength scale and the corresponding criterion across the four samples. To evaluate the discriminant validity of the strength scales, the number of other criterion scales that correlated more highly with the strength scale than the expected criterion was summed across the samples. For example, the entry of 6 for the VIA-IS-R Gratitude scale means that across the four samples there were six criterion measures that correlated more highly with the VIA-IS-R scale than the expected criterion. A value of 0 therefore means no unexpected strength-criterion correlation was larger than the expected one in any sample. At the bottom of the columns are means as well as the number of strength scales where an unexpected correlation was larger than the expected correlation in at least one sample. The columns for the VIA-IS-R are based on all four samples. All other columns exclude the GC Sample.

Table 6. External validity analyses across samples.

Convergent validity results were quite consistent across the VIA-IS-R/M/P inventories, with mean correlations around .55 for all three versions and a minimum mean value of .35. Prior research suggests correlations of .30 or higher represent relatively large correlations in practice (e.g., Brydges, Citation2019; Hemphill, Citation2003), while McGrath et al. (Citation2020) used several lines of evidence to justify the conclusion that correlations ≥ .60 provide evidence of “likely redundancy” (p. 130). Mean correlations for the GACS-72 and -24 were somewhat smaller. Discriminant validity results were generally good, with some problems noted for Curiosity, Gratitude, and Teamwork. Interestingly, for these two strengths every problem with discriminant validity except six (five for Gratitude and one for Teamwork) involved the MC Sample; the reason for this finding is unknown.

There were several interesting results for the GC Sample given concerns that have been raised about the virtue placement for some of the strengths in the original model. If Hope was considered a Courage strength, as previous research has suggested, the discriminant validity value for the Hope scale would drop to 0. In contrast, the Humor scale unexpectedly correlated best with the Wisdom item on the virtues scale. Research mentioned earlier (Ruch et al., Citation2020) suggested Leadership is interstitial across the virtues Wisdom and Knowledge, Courage, and its hypothesized placement as a Judgment variable. In the GC sample, the correlations of Leadership with Wisdom and Knowledge and Courage both exceeded the expected correlation between Leadership and Justice. Similarly, Honesty correlated more with self-perceived Justice than its expected correlate Courage, as previous research has suggested. It also correlated more with Temperance, which was a new finding.

Mean convergent validity correlations were lower for the two versions of the GACS. Comparing internal reliability and external validity results for the GACS-72/24 suggests a classic example of the attenuation paradox (Loevinger, Citation1954; Smith et al., Citation2000), in which the use of more similar items (often in the process of developing a shorter scale) can optimize reliability and unidimensionality while sacrificing content coverage, thereby reducing validity coefficients. However, all means were still .29 or greater, with the Humility and Teamwork scales demonstrating the poorest convergent validity. Consistent with findings for the VIA-IS-R/M/P scales, for these two strengths 33 of 34 discriminant validity failures involved the MC Sample.

Discussion

Given the number of analyses, a summary of the findings is worthwhile:

  1. The VIA-IS-R and GACS-72 scales were the most reliable across the different inventories, with mean values approaching or surpassing .80. This is not surprising given the former has the most items per scale and the latter has the most consistent items in content. The values for the shorter instruments were weaker. Most still exceeded .70 on average, though users may consider the values for Humility and Judgment on the VIA-IS-M and VIA-IS-P less than desirable or even unacceptable. Test-retest reliability results were similar.

  2. Alternate versions reliability across the VIA-IS-R/M/P inventories was quite good. The results in suggest that, for example, a study comparing pretest to ‌posttest data on strengths and virtues might use different short forms for the two administrations as a means of reducing practice effects. However, this finding can be faulted in that it was based on generating short version scores from administration of the VIA-IS-R. A better strategy would involve administering the three as discrete instruments, though doing so would probably prove burdensome to the participants. A future study might employ an administration strategy involving planned missing data used as in the case of the GC Sample to make such an analysis more practical.

  3. It was possible to achieve unidimensionality after freeing a relatively small set of error covariances. None were needed for the GACS-72 scales, and no more than one for any of the VIA-IS-M and -P scale. These findings are particularly surprising given that unidimensionality was one of only several factors considered in item selection (McGrath, Citation2019).

  4. Using the final unidimensional models, most tests of measurement invariance indicated scalar invariance was achieved in a majority of the samples. There were three scales for which this standard was not met in the case of both gender and age on the VIA-IS-R: Bravery, Curiosity, and Humility. Judgment also demonstrated problems with gender on the VIA-IS-R and VIA-IS-M, the two instruments with reverse-keyed items. The majority of samples did achieve metric invariance in all cases except the Humility scale, suggesting covariation comparisons between demographic subgroups can still be meaningful. The VIA-IS-P and GACS-72 achieved scalar invariance in all cases.

  5. Evidence for convergent and discriminant validity was generally good, though problems with the latter were noted for the Curiosity, Gratitude, and Teamwork scales. Convergent validity for the GACS-24 was on average only slightly lower than that for the GACS-72, and discriminant validity results were slightly better. These findings have to be interpreted with particular caution given that none of the criterion measures used for these analyses has been validated in their own right. Unfortunately, most studies that involve measures of the 24 strengths focus exclusively on well-being and flourishing as an outcome (e.g., Blanchard et al., Citation2020; Proyer et al., Citation2011), which offers no basis for judging discriminant validity at all.

Of course, this study has its limitations. Some of the statistics gathered for this article were reviewed in the VD sample in the process of selecting the items for the VIA-IS-R/M/P. However, item selection was based on a balancing of considerations, including some such as item response theory results that played no role in the present study. The extent to which the resulting scales meet all standards for adequate validity is therefore worth evaluating. These statistics also played no role in the development of the GACS-72/24 items.

Second, two of the samples were recruited through the VIA website. These are individuals who demonstrate a preexisting interest in the VIA Classification. Their decision to complete a lengthy questionnaire in return for personal feedback, as well as their willingness to respond to an extensive battery of additional items for research purposes, carves them out as specialized samples. These concerns provided the rationale for paying two samples of adults to participate in the research, and in one case to attempt a representative sampling of population demographic categories (McGrath et al., Citation2022; McGrath & Wallace, Citation2021). Of course, the types of individuals accessible through online panels are hardly representative of the population in general. That said, the present samples may well be no poorer than many of the adult samples used in personality research.

Given the variety of instruments discussed here, it is worth concluding with a few recommendations for their use. If the primary goal is to maximize both reliability and external validity (a common preference in practice), the VIA-IS-R is probably the best choice, though research with a broader array of outcome measures would be helpful for verifying this recommendation. It offers the greatest breadth of item contents and demonstrates consistently desirable psychometric qualities.

If length is the paramount consideration, the GACS-72, or even the GACS-24, provides a credible choice. If brevity is a key consideration while still wanting to achieve better content coverage, the two short forms should be considered. The VIA-IS-M in general performed more poorly than the other two versions of that inventory. However, as noted earlier, some test users object to unidirectional item keying because of concerns about yea-saying and nay-saying response sets. The better option for these users is the VIA-IS-R, but when that instrument is impractical because of its length the VIA-IS-M offers a reasonably acceptable alternative.

An implicit theme of this article has been that no instrument is or can be perfect. There will always be tensions between the varying considerations for evaluating psychological measures: maximizing unidimensionality, variation in item difficulties, content coverage, convergent validity, and discriminant validity among others. number of standards. That said, the VIA-IS-P may well be the best single choice for balancing issues of structural validity, external validity, content coverage, cognitive load, and length. The English language version of the instrument currently administered by default on the VIA Institute website is the VIA-IS-P. As translations of this instrument are completed (Bates & McGrath, n.d.), they will be substituted for translations of the VIA-120.

It should be noted that other multi-item measures of the VIA strengths are available for adults. Ng and associates developed both an alternate measure of the 24 strengths called the Comprehensive Inventory of Virtuous Instantiations of Character, or CIVIC (Ng et al., Citation2018), as well as a short form of the original VIA-IS (Ng et al., Citation2017). In both cases, they emphasized scale unidimensionality as the primary criterion for item selection. They have also introduced a forced choice version of the CIVIC in an attempt to reduce socially desirable responding about character strengths (Ng et al., Citation2021). Several inventories measuring the 24 character strengths have been developed using items from the International Personality Item Pool (Goldberg et al., Citation2006), a freely available database of over 3,000 personality items. One version was developed with the intention of optimizing statistics from item response theory analyses (du Plessis & de Bruin, Citation2015), while another focused primarily on unidimensionality and invariance across English and German versions (Partsch et al., Citation2022). At least five other measures using a single item for each character strength have been developed besides the GACS-24 (Cosentino & Solano, Citation2012; Furnham & Lester, Citation2012; Ruch et al., Citation2014; Vanhove et al., Citation2016; Vie et al., Citation2016), two of which were developed for use in military settings. Ruch et al. (Citation2020) also described several measures of the six original virtues specifically and virtuousness in general.

It is also worth noting that several measures have been developed for purposes other than measuring the trait level of the strengths. The Overuse, Underuse, Optimal-Use of Character Strengths (Freidlin et al., Citation2017) builds on the Aristotelian concept of the golden mean for virtues by asking the respondent to indicate what percent of the time they use each strength in a manner that suggests insufficient use, excessive use of the strength, or a desirable level of use. The Partners Strengths Questionnaire (Kashdan et al., Citation2018) allows measurement of appreciation for the strengths of one’s partner. The Strengths Use Scale (Govindji & Linley, Citation2007) was developed for purposes of coaching to gauge the extent to which the respondent uses each of the strengths in daily life. Lamson and McGrath (Citation2021) have even developed a text analysis tool for the 24 strengths.

All of these instruments are free to use. As noted earlier, the VIA Institute requires permission to use the VIA-IS-R/M/P, but this is primarily to track research, and no reasonable request is rejected. There is clearly no shortage of alternatives to the VIA Assessment Suite tools, and some of these instruments serve specific purposes not addressed by the measures in the suite. The VIA-IS-R/M/P and GACS-72/24 are distinct from these others in the existence of five samples explicitly collected for purposes of validation, including latent structural analyses and scale-specific correlates used for external validation. These features make them the most thoroughly vetted measures of the character strengths available. The Institute is also actively involved in translation of these measures into other languages, with 13 completed to date.

Author’s note

Robert McGrath is a Senior Scientist for the VIA Institute on Character, the copyright holder for the instruments examined in this manuscript.

Acknowledgments

I am grateful to the many coauthors who helped make this work possible. The data and syntax used in analyses described in this article are available upon request to the author.

Additional information

Funding

The research summarized in this article was funded in part by the VIA Institute on Character.

Notes

1 The Assessment Suite includes a third measure called the Signature Strengths Survey (SSS; McGrath, Citation2019), which uses a checklist that allows respondents to self-identify signature strengths. This instrument is unusual in that it generates a binary outcome for each strength, it is infrequently used, and has been less extensively vetted than the other two. For these reasons, it was omitted from this review of the instruments that use more typical item formats. The statements made in this paragraph about the GACS-72/24 apply to the SSS as well.

References

  • Arbenz, G. C., Gander, F., & Ruch, W. (in press). Breadth, polarity, and emergence of character strengths and their relevance for assessment. Journal of Positive Psychology. https://doi.org/10.1080/17439760.2021.2018026
  • Bates, F. C., & McGrath, R. E. (n.d.). Employing best practices when translating positive psychology questionnaires: The VIA Assessment Suite for Adults. VIA Institute on Character. https://via-assets.global.ssl.fastly.net/76d1ea39-a4eb-4270-b9dc-899653415f8f/assets/TranslationofVIAAdultAssessmentSuite.pdf
  • Blanchard, T., Kerbeykian, T., & McGrath, R. E. (2020). Why are signature strengths and well-being related? Tests of multiple hypotheses. Journal of Happiness Studies, 21(6), 2095–2114. https://doi.org/10.1007/s10902-019-00170-2
  • Broughton, R. (1984). A prototype strategy for construction of personality scales. Journal of Personality and Social Psychology, 47(6), 1334–1346. https://doi.org/10.1037/0022-3514.47.6.1334
  • Brydges, C. R. (2019). Effect size guidelines, sample size calculations, and statistical power in gerontology. Innovation in Aging, 3(4), igz036. https://doi.org/10.1093/geroni/igz036
  • Buss, D. M., & Craik, K. H. (1983). The act frequency approach to personality. Psychological Review, 90(2), 105–126. https://doi.org/10.1037/0033-295X.90.2.105
  • Cortina, J. M. (1993). What is coefficient alpha? An examination of theory and applications. Journal of Applied Psychology, 78(1), 98–104. https://doi.org/10.1037/0021-9010.78.1.98
  • Cosentino, A. C., & Solano, A. C. (2012). Character strengths: A study of Argentinean soldiers. The Spanish Journal of Psychology, 15(1), 199–215. https://doi.org/10.5209/rev_SJOP.2012.v15.n1.37310
  • Dahlsgaard, K., Peterson, C., & Seligman, M. E. P. (2005). Shared virtue: The convergence of valued human strengths across culture and history. Review of General Psychology, 9(3), 203–213. https://doi.org/10.1037/1089-2680.9.3.203
  • Du Plessis, G. A., & de Bruin, G. P. (2015). Using Rasch modelling to examine the International Personality Item Pool (IPIP) Values In Action (VIA) measure of character strengths. Journal of Psychology in Africa, 25(6), 512–521. https://doi.org/10.1080/14330237.2015.1124603
  • Feraco, T., Casali, N., & Meneghetti, C. (2022). Do strengths converge into virtues? An item-, virtue-, and scale-level analysis of the Italian Values in Action Inventory of Strengths-120. Journal of Personality Assessment, 104(3), 395–407. https://doi.org/10.1080/00223891.2021.1934481
  • Flake, J. K., Pek, J., & Hehman, E. (2017). Construct validation in social and personality research: Current practice and recommendations. Social Psychological and Personality Science, 8(4), 370–378. https://doi.org/10.1177/1948550617693063
  • Freidlin, P., Littman-Ovadia, H., & Niemiec, R. M. (2017). Positive psychopathology: Social anxiety via character strengths underuse and overuse. Personality and Individual Differences, 108, 50–54. https://doi.org/10.1016/j.paid.2016.12.003
  • Furnham, A., & Lester, D. (2012). The development of a short measure of character strength. European Journal of Psychological Assessment, 28(2), 95–101. https://doi.org/10.1027/1015-5759/a000096
  • Goldberg, L. R., Johnson, J. A., Eber, H. W., Hogan, R., Ashton, M. C., Cloninger, C. R., & Gough, H. G. (2006). The International Personality Item Pool and the future of public-domain personality measures. Journal of Research in Personality, 40(1), 84–96. https://doi.org/10.1016/j.jrp.2005.08.007
  • Govindji, R., & Linley, P. A. (2007). Strengths use, self-concordance and well-being: Implication for strengths coaching and coaching psychologists. International Coaching Psychology Review, 2(2), 143–153.
  • Gustems-Carnicer, J., & Calderón, C. (2016). Virtues and character strengths related to approach coping strategies of college students. Social Psychology of Education, 19(1), 77–95. https://doi.org/10.1007/s11218-015-9305-y
  • Hemphill, J. F. (2003). Interpreting the magnitudes of correlation coefficients. American Psychologist, 58(1), 78–79. https://doi.org/10.1037/0003-066X.58.1.78
  • Hopwood, C. J., & Donnellan, M. B. (2010). How should the internal structure of personality inventories be evaluated? Personality and Social Psychology Review 14(3), 332–346. https://doi.org/10.1177/1088868310361240
  • Hu, L-t., & Bentler, P. M. (1999). Cutoff criteria for fit indexes in covariance structure analysis: Conventional criteria versus new alternatives. Structural Equation Modeling: A Multidisciplinary Journal, 6(1), 1–55. https://doi.org/10.1080/10705519909540118
  • Hussey, I., & Hughes, S. (2020). Hidden invalidity among 15 commonly used measures in social and personality psychology. Advances in Methods and Practices in Psychological Science, 3(2), 166–184. https://doi.org/10.1177/2515245919882903
  • Kashdan, T. B., Blalock, D. V., Young, K. C., Machell, K. A., Monfort, S. S., McKnight, P. E., & Ferssizidis, P. (2018). Personality strengths in romantic relationships: Measuring perceptions of benefits and costs and their impact on personal and relational well-being. Psychological Assessment, 30(2), 241–258. https://doi.org/10.1037/pas0000464
  • Kelley, K. (2020). The MBESS R Package: R package version 4.8.0. https://CRAN.R-project.org/package=MBESS
  • Lamson, W. R. J., & McGrath, R. E. (2021). Speaking of character: Character strength references in movies and presidential nomination speeches. Journal of Positive Psychology, 16(2), 218–227. https://doi.org/10.1080/17439760.2019.1689415
  • Linley, P. A., Nielsen, K. M., Gillett, R., & Biswas-Diener, R. (2010). Using signature strengths in pursuit of goals: Effects on goal progress, need satisfaction, and well-being, and implications for coaching psychologists. International Coaching Psychology Review, 5(1), 6–15.
  • Loevinger, J. (1954). The attenuation paradox in test theory. Psychological Bulletin, 51(5), 493–504. https://doi.org/10.1037/h0058543
  • McDonald, R. P. (1999). Test theory: A unified treatment. Erlbaum.
  • McGrath, R. E. (2015). Integrating psychological and cultural perspectives on virtue: The hierarchical structure of character strengths. Journal of Positive Psychology, 10(5), 407–424. https://doi.org/10.1080/17439760.2014.994222
  • McGrath, R. E. (2019). Technical report: The VIA assessment suite for adults: Development and initial evaluation (rev. ed.). VIA Institute on Character. https://evada-assets.global.ssl.fastly.net/76d1ea39-a4eb-4270-b9dc-899653415f8f/assets/Technical%20Report%20Revised%20Edition%202019_1.pdf
  • McGrath, R. E. (2021). Darwin meets Aristotle: Fundamental virtues as contributors to addressing current cultural tensions. Journal of Positive Psychology, 16(4), 431–445. https://doi.org/10.1080/17439760.2020.1752781
  • McGrath, R. E., Brown, M., Westrich, B., & Han, H. (2022). Representative sampling of the VIA assessment suite for adults. Journal of Personality Assessment, 104(3), 380–394. https://doi.org/10.1080/00223891.2021.1955692
  • McGrath, R. E., Greenberg, M. J., & Hall-Simmonds, A. (2018). Scarecrow, Tin Woodsman, and cowardly lion: The three-factor model of virtue. Journal of Positive Psychology, 13(4), 373–392. https://doi.org/10.1080/17439760.2017.1326518
  • McGrath, R. E., Hall-Simmonds, A., & Goldberg, L. R. (2020). Are measures of character and personality distinct? Evidence from observed-score and true-score analyses. Assessment, 27(1), 117–135. https://doi.org/10.1177/1073191117738047
  • McGrath, R. E., Mitchell, M., Kim, B. H., & Hough, L. (2010). Evidence for response bias as a source of error variance in applied assessment. Psychological Bulletin, 136(3), 450–470. https://doi.org/10.1037/a0019216
  • McGrath, R. E., & Wallace, N. (2021). Cross-validation of the VIA Inventory of strengths-revised and its short forms. Journal of Personality Assessment, 103(1), 120–131. https://doi.org/10.1080/00223891.2019.1705465
  • McNeish, D. (2018). Thanks coefficient alpha, we’ll take it from here. Psychological Methods, 23(3), 412–433. https://doi.org/10.1037/met0000144
  • Merritt, S., Huber, K., & Bartkoski, T. (2019). Application of signature strengths at work: A dual-level analysis. Journal of Positive Psychology, 14(1), 113–124. https://doi.org/10.1080/17439760.2018.1519589
  • Ng, V., Cao, M., Marsh, H. W., Tay, L., & Seligman, M. E. P. (2017). The factor structure of the Values in Action Inventory of Strengths (VIA-IS): An item-level exploratory structural equation modeling (ESEM) bifactor analysis. Psychological Assessment, 29(8), 1053–1058. https://doi.org/10.1037/pas0000396
  • Ng, V., Lee, P., Ho, M.-H R., Kuykendall, L., Stark, S., & Tay, L. (2021). The development and validation of a multidimensional forced-choice format character measure: Testing the Thurstonian IRT approach. Journal of Personality Assessment, 103(2), 224–237. https://doi.org/10.1080/00223891.2020.1739056
  • Ng, V., Tay, L., & Kuykendall, L. (2018). The development and validation of a measure of character: The CIVIC. Journal of Positive Psychology, 13(4), 346–372. https://doi.org/10.1080/17439760.2017.1291850
  • Park, N., & Peterson, C. (2006). Moral competence and character strengths among adolescents: The development and validation of the Values in Action Inventory of Strengths for Youth. Journal of Adolescence, 29(6), 891–909. https://doi.org/10.1016/j.adolescence.2006.04.011
  • Partsch, M. V., Bluemke, M., & Lechner, C. M. (2022). Revisiting the hierarchical structure of the 24 VIA character strengths: Three global dimensions may suffice to capture their essence. European Journal of Personality, 36(5), 825–845. https://doi.org/10.1177/08902070211017760
  • Peterson, C., & Seligman, M. E. P. (2004). Character strengths and virtues: A classification and handbook. American Psychological Association/Oxford University Press.
  • Peterson, R. A., Kim, Y., & Choi, B. (2020). A meta-analysis of construct reliability indices and measurement model fit metrics. Methodology, 16(3), 208–223. https://doi.org/10.5964/meth.2797
  • Proyer, R. T., Gander, W., Wyss, T., & Ruch, W. (2011). The relation of character strengths to past, present, and future life satisfaction among German-speaking women. Applied Psychology: Health and Well-Being, 3(3), 370–384. https://doi.org/10.1111/j.1758-0854.2011.01060.x
  • R Core Team. (2021). R: A language and environment for statistical computing. R Foundation for Statistical Computing. https://www.R-project.org
  • Revelle, W. (2020). psych: Procedures for personality and psychological research (Version 2.0.12). Northwestern University. https://CRAN.R-project.org/package=psych
  • Rosseel, Y. (2012). lavaan: An R package for structural equation modeling. Journal of Statistical Software, 48(2), 1–36. https://doi.org/10.18637/jss.v048.i02
  • Ruch, W., Heintz, S., & Wagner, L. (2020). Co-occurrence patterns of character strengths and measured core virtues in German-speaking adults. Frontiers in Psychology, 11. Article 599094. https://doi.org/10.3389/fpsyg.2020.599094
  • Ruch, W., Martinez-Marti, M. L., Proyer, R. T., & Harzer, C. (2014). The Character Strengths Rating Form (CSRF): Development and initial assessment of a 24-item rating scale to assess character strengths. Personality and Individual Differences, 68, 53–58. https://doi.org/10.1016/j.paid.2014.03.042
  • Shrout, P. E. (1998). Measurement reliability and agreement in psychiatry. Statistical Methods in Medical Research, 7(3), 301–317. https://doi.org/10.1177/096228029800700306
  • Sijtsma, K. (2009). On the use, the misuse, and the very limited usefulness of Cronbach’s alpha. Psychometrika, 74(1), 107–120. https://doi.org/10.1007/s11336-008-9101-0
  • Smith, G. T., McCarthy, D. M., & Anderson, K. G. (2000). On the sins of short-form development. Psychological Assessment, 12(1), 102–111. https://doi.org/10.1037/1040-3590.12.1.102
  • Suldo, S. M., Hearon, B. V., Bander, B., McCullough, M., Garofano, J., Roth, R. A., & Tan, S. Y. (2015). Increasing elementary school students’ subjective well-being through a classwide positive psychology intervention: Results of a pilot study. Contemporary School Psychology, 19(4), 300–311. https://doi.org/10.1007/s40688-015-0061-y
  • Uliaszek, A. A., Rashid, T., & Zarowsky, Z. (2022). The role of signature strengths in treatment outcome: Initial results from a large and diverse university sample. Journal of Contemporary Psychotherapy, 52(1), 15–21. https://doi.org/10.1007/s10879-021-09523-6
  • Vanhove, A. J., Harms, P. D., & DeSimone, J. A. (2016). The Abbreviated Character Strengths Test (ACST): A preliminary assessment of test validity. Journal of Personality Assessment, 98(5), 536–544. https://doi.org/10.1080/00223891.2016.1148044
  • Vie, L. L., Scheier, L. M., Lester, P. B., & Seligman, M. E. P. (2016). Initial validation of the U.S. Army Global Assessment Tool. Military Psychology, 28(6), 468–487. https://doi.org/10.1037/mil0000141
  • Vylobkova, V., Heintz, S., Gander, F., Wagner, L., & Ruch, W. (in press). Convergence and psychometric properties of character strengths measures: The VIA-IS and the VIA-IS-R. Journal of Personality Assessment. https://doi.org/10.1080/00223891.2022.2044342
  • Wanous, J. P., & Hudy, M. J. (2001). Single-item reliability: A replication and extension. Organizational Research Methods, 4(4), 361–375. https://doi.org/10.1177/109442810144003