2,523
Views
9
CrossRef citations to date
0
Altmetric
Articles

Unpacking domain-specific achievement motivation: the role of contextualising items for test-criterion correlations

, , &
Pages 501-525 | Received 11 Jan 2019, Accepted 05 Jan 2020, Published online: 14 Feb 2020

Abstract

Achievement motivation scores on the domain-specific level are better predictors of domain-matching scholastic performance than scores of general achievement motivation measures. Although there is research on domain-specific motivational measures, it is still unknown where this higher predictive power originates from. To address this, 715 students in secondary school answered questionnaires on general and domain-specific achievement motivation, domain-specific self-concept, and domain-specific self-esteem in two different studies. The first study was designed to disentangle the variance components in general and domain-specific achievement motivation in order to delineate hypotheses regarding potential drivers for the predictive power of domain-specific achievement motivation. The findings implied a strong role for a shared method factor. To explore the nature of this method factor, domain-specific self-concept/-esteem were focussed to establish discriminant validity evidence in a second study. The results indicate that the additional domain-specific variance can, in large parts, be explained by self-concept and self-esteem on domain-specific level.

The interplay between personal goals, self-beliefs, motives, and scholastic achievement is of growing interest to many researchers and practitioners, especially since the increasing recognition that motivational constructs contribute to the prediction of scholastic performance incrementally to intelligence and broader personality traits (Steinmayr & Spinath, Citation2007, Citation2009). Test scores from achievement motivation measures have been shown to have robust and replicable test-criterion correlations with criteria of scholastic performance (Richardson, Abraham, & Bond, Citation2012). Prior research suggested that scores from domain-specific achievement motivation measures outperform scores from general achievement motivation measures in terms of test-criterion correlations (Steinmayr & Spinath, Citation2009). Thus, through contextualising items by, for example, adding a phrase like ‘…in math’, achievement motivation scores correlate stronger with domain-specific performance. The present studies add to this literature by applying variance decomposition models to generally and domain-specifically phrased achievement motivation measures. These analyses allow to quantify trait variance and method variance due to contextualising, and to investigate the contribution of each variance source when predicting scholastic performance. Finally, we test whether the variance based on contextualising such achievement motivation items can be discriminated from other, achievement relevant constructs such as self-concept and self-esteem.

Domain-specific achievement motivation and scholastic performance

In their review, Murphy and Alexander (Citation2000, p. 41) summarised that research on motivation develops ‘toward a more domain-specific viewpoint’. This, along with the increasing body of literature utilising domain-specific scales, was the rationale for this research project. As stated, a growing number of studies focus on domain-specific achievement motivation when predicting academic performance on domain level (Green, Martin, & Marsh, Citation2007; Wigfield, Citation1997; Wigfield, Guthrie, Tonks, & Perencevich, Citation2004). Domain-specific measures of achievement motivation play an important role in predicting interindividual differences in scholastic performance. Interestingly, the use of such contextualised items stands in contrast to traditional definitions for many of the constructs contextualisation is applied to. For example, Murray defined achievement motivation in a way that suggests that it acts the same way, regardless of the situation (Steinmayr & Spinath, Citation2009). However, research on the situational consistency of traits shows that even relatively stable traits, such as the Big Five, underlie situational variations (Deinzer et al., Citation1995). Moreover, adding contextualising phrases potentially changes the frame of reference (Schmit, Ryan, Stierwalt, & Powell, Citation1995; Shaffer & Postlethwaite, Citation2012) and thereby the answer process (Krosnick, Citation1999; Ziegler, Citation2011). Another relevant aspect is that contextualising changes the level of abstraction causing scale scores from contextualised items to be more symmetrical to many outcomes than scale scores from domain-general items (Kretzschmar, Spengler, Schubert, Steinmayr, & Ziegler, Citation2018). This is why, despite the inherent violation of underlying definitions, researchers before us have called for future studies which ‘… should compare the predictive power of the constructs assessed at the same level of generality and include further motivational constructs … . A possible design should cover a complete crossover of the investigated motivational concepts and level of generality, i.e. need for achievement, ability self-concepts, values, and goals would be assessed with regard to motivation in general, at school, and for different subjects. Even though such a design would not necessarily assess each motivational construct in accordance with its theoretical foundation, it would allow for a more straightforward comparison of the predictive power of the concepts’ (Steinmayr & Spinath, Citation2009, p. 89). Wigfield and Cambria (Citation2010), in a review paper, also summarised that the question of domain specificity of motivational measures is clear for some constructs (e.g. self-efficacy) and less clear for others (e.g. task orientation) which in their opinion is mainly due to the lack of domain-specifically phrased measures for some of the constructs. Following the call by Steinmayr and Spinath (Citation2009) and the other mentioned ideas, understanding the variance makeup of contextualised measures is an important step. Although the body of research on the predictive power of domain-specific motivational measures is growing, the only explanation for their relatively better test-criterion correlations so far is the level of symmetry between predictor and criteria ( Brunswik, Citation1955; also see Bandura, Citation1997; Mischel, Citation1977). The key goal of the present research was to explore the change in score variance introduced by contextualising items by decomposing the variance into its different components, that is variance due to the trait being measured, variance due to contextualising (method variance), and measurement error.

Different theoretical approaches to achievement motivation in the school context

Among the variety of achievement-related motivational constructs, we concentrate on the need for achievement and goal-orientations (Murphy & Alexander, Citation2000). Clearly, the construct space in this field is much broader. However, we had to select a few constructs to pay tribute to the practical limitations of collecting data in schools. The constructs were therefore selected based on prior work showing their relevance for performance in different school subjects (Steinmayr & Spinath, Citation2007, Citation2009; Ziegler, Knogler, & Bühner, Citation2009; Ziegler, Schmidt-Atzert, Bühner, & Krumm, Citation2007). Accordingly, our results are limited to the selection made and future research needs to include a broader range of constructs.

Need for achievement and scholastic performance

Murray (Citation1938) defined need for achievement as a drive to reach internal or external goals and called it a basic human need. McClelland, Atkinson, Clark, and Lowell (Citation1953) differentiated the need for achievement into two achievement motives: Hope for success and fear of failure. They also characterised achievement situations by an emotional conflict in which hope for success and fear of failure have to be balanced. These motives are believed to arise before the actual achievement situation thereby determining the direction, intensity, and quality of achievement-related behaviour. While McClelland was highly skeptical of self-report measures (McClelland, Koestner, & Weinberger, Citation1989), over the years self-reports have become the standard means to capture both motives (e.g. Achievement Motivation Questionnaire by Nygård & Gjesme, Citation1973). McClelland et al. (Citation1989) hypothesised that such self-reports of achievement motivation might be influenced by self-concept. Sparfeldt and Rost (Citation2011) explored how well general versus domain-specific variants of hope for success and fear of failure predict grades in four subjects (mathematics, German, physics, and English). They found that domain-specific achievement motives showed stronger test-criterion correlations with their respective grades than general hope for success or fear of failure scores.

Goal theories and scholastic performance

Early achievement motivation theories assumed that capacity or ability could be seen based on mastery of a subject or based on performance comparisons (Nicholls, Citation1984). These thoughts were developed into the concepts of mastery and performance goals (Ames & Archer, Citation1987). While mastery orientation implies the desire of a person to gain competence, a performance-oriented person aims to outperform others or an external standard like grades. This dichotomous framework was further developed towards the 2 × 2 achievement goal framework in which a mastery-performance dimension is crossed with an approach-avoidance dimension (Elliot & McGregor, Citation2001). Approach goals motivate a person to approach situations in which competence can be shown, while avoidance goals make a person avoid such situations. The 2 × 2 framework thus postulates mastery-approach and mastery-avoidance, performance-approach and performance-avoidance goals. Its focus is on emotions experienced during or after achievement situations. Those can be assessed independently with the Achievement Goal Questionnaire (AGQ; Elliot & McGregor, Citation2001). The psychometric properties of the AGQ and the theoretical plausibility of the 2 × 2 framework have been shown in many studies (e.g. Elliot & Murayama, Citation2008). It has to be noted, though, that the mastery-avoidance scale often shows suboptimal psychometric properties, especially internal consistencies (Finney, Pieper, & Barron, Citation2004). Wirthwein, Sparfeldt, Pinquart, Wegerer, and Steinmayr, (Citation2013) conducted a meta-analysis on the relations between goal-orientations and academic achievement. They found that matching specificity is a significant moderator, which again attests to the better test-criterion correlation of domain-specific scale scores.

Variance decomposition as a way to analyse test-criterion correlations

Despite the consistent empirical evidence showing the higher test-criterion correlations for scores from domain-specifically phrased measures compared to scores from generally phrased measures, explanations for this effect are scarce. We propose that there must be different variance sources within those scores, which show differential correlations with scholastic performance criteria. To identify such specific relations, it is helpful to first compare generally and domain-specifically phrased scales in terms of their variance components (Krumm, Hüffmeier, & Lievens, Citation2019). Items of these scales should be identical, except for the addition of ‘in math’, to make sure that they share the same content variance, which pertains to the actual trait measured (e.g. hope for success) and only differs in the methodological, context-specifying addition (i.e. generally versus in math).

When answering a contextualised item, test-takers are supposedly accessing a somewhat different construct compared to the generally phrased items. Current theory suggests that this construct is a domain-specific facet of the general construct. However, this explanation has not been formally tested. Nevertheless, modelling the different variance sources (i.e., trait variance and variance due to the added phrase) would allow two things: First, the actual amount of variance due to contextualising could be estimated. Second, the variance components could be correlated with other measures, to shed light on their role in predicting criteria like scholastic performance.

One approach that seems suitable for this purpose is so called Correlated-Trait-Correlated-Method minus One model (CTC(M-1)) which was developed by Eid et al. (e.g. Eid et al., Citation2008). These models serve the purpose of decomposing scale variance into trait, method, and error variance into separate latent variables. For our particular case, this implies that both the general and the specific item, for example capturing hope for success, should be loaded by a latent trait factor hope for success, but the specific item would additionally be loaded by a corresponding latent domain method factor. Each item would have its individual latent error. Based on the loadings with the latent trait factor, a consistency coefficient can be estimated which reflects the amount of systematic variance within each measure due to the common trait in question (the variance due to the trait is divided by the sum of this variance and the variance due to the method). One important aspect of CTC(M-1) models is that they require choosing one method (e.g. generally or domain-specifically phrased) as the reference method (thus, the ‘minus 1’ in the model name). The variance due to this method is not specified as a latent variable in the models. One statistical advantage is that the models have fewer convergence problems (Eid et al., Citation2008). A CTC(M-1) model allows to estimate a domain-specific method factor, which is interpretable as the deviation of the domain-specific scale from the score which would be expected based on the reference method. CTC(M-1) models further allow estimating the amount of variance in each scale due to this method variable (method specificity: variance due to the method divided by the sum of this variance and the variance due to the trait). Moreover, it is possible to use the different variance components (trait versus method factor) as correlates of criteria. Thus, CTC(M-1) models inform us about the relative proportion of variance shared between the general and the domain-specific scale versions, in addition to non-shared variance (i.e. method variance due to contextualising the scales).

We conducted two studies, each using a systematic experimental item variation. We used items which only differed in their specificity (general versus in various specific domains) in combination with CTC(M-1) modelling. Whereas Study 1 focussed on the variance decomposition and the role each variance component has for test-criterion correlations, Study 2 explored possible correlates of the latent method factor to test its discriminant validity.

Study 1

Despite the increased recognition of domain-specific achievement motivation measures when predicting scholastic performance, it remains unclear what the specific role of variance due to contextualising items is with regard to test-criterion correlations. With this study we want to add to the existing literature by addressing four questions: First, are measure specific method factors necessary? Second, does the trait variance within general and domain-specific achievement measures show convergent validity? Third, what are the relative proportions each variance source has in general and domain-specifically phrased measures? Fourth, what is the role of each variance source for test-criterion correlations?

Method of Study 1

Sample and power

We collected data from 325 students from two different schools (174 girls), aged 14.32 on average (SD = 0.92). The majority (n=174) attended German ‘Hauptschule’, a school which offers lower secondary education, n=151 attended German Gymnasium’, a school that offers higher secondary education in preparation for university. Students also differed in their grades with 194 students in 8th grade and 131 in the 9th grade.

Procedure and measures

Procedure

All data were collected via student self-report, during regular class hours on a voluntary basis after having received written consent from their parents. Most questionnaires needed to be adapted to get a general and a domain-specific version (see Appendix 2). Thus, a test battery was generated that included general versions of all questionnaires used as well as math specific versions of the same questionnaires. The latter was created by adding ‘in math’ or ‘in math class’ to general items. Descriptive statistics and internal consistencies for all measures can be found in Appendix 1. All internal consistency estimates were satisfactory with the exception of mastery-avoidance. As the modelling approach chosen uses latent variables and corrects for attenuation effects between latent variables, the score was used nevertheless. Moreover, there is empirical evidence suggesting that internal consistency estimates have little influence on test-criterion correlations focussed here (McCrae, Kurtz, Yamagata, & Terracciano, Citation2011).

Measures

We used adapted versions of the approach versus avoidance by mastery versus performance motivation scales of Elliot and McGregor’s Achievement Goal Questionnaire (2001) and the revised German Achievement Motives Scale (AMS-R; Lang & Fries, Citation2006; cf. Nygård & Gjesme, Citation1973, see Appendix 2 for further details).

Scholastic performance

School grades in mathematics, German, and physics were reported by the students and used as criteria of scholastic performance. Grades range from 1, the best, to 6, the worst grade, with 5 and 6 indicating insufficient performance. We inverted the grades, in order to facilitate interpretation.

Statistical analyses

Models

R (R Core Team, Citation2016) and Mplus 8.1 (Muthén & Muthén, Citation1998–2012) were used for all analyses. Robust maximum-likelihood estimation (MLR) plus type two-level complex with school type as stratification and school class as cluster variables were used to analyse the structural equation models. These specifications help to deal with violations of the multivariate normal distribution as well as with the nested data structure due to the two school types and students being in different classes. Model fit was judged on the basis of the χ2-test, the Standardised Root Mean Square Residual (SRMR), and the Root Mean Square Error of Approximation (RMSEA) as recommended by Hu and Bentler (Citation1999), and on the Comparative Fit Index (CFI) as advised by Beauducel and Wittmann (Citation2005). We used the following cut-off criteria for assuming good model fit: the SRMR should be smaller than 0.11 in combination with an RMSEA of less than or equal to 0.08 (n>250), and a CFI value above 0.95.

The first CTC(M-1) model (Model 1) specified six correlated latent trait factors corresponding to the different theoretical approaches to achievement motivation: mastery-approach (MAP), mastery-avoidance (MAV), performance-approach (PAP), performance-avoidance (PAV), hope for success (HS), and fear of failure (FF) (). Scores for each instrument and each level of globality (domain-specifically versus generally phrased) were also loaded by a method factor. In order to create a CTC(M-1) model, not all method factors were specified, though. As we considered the test family as a method, we decided to use the AMS as a reference method. Therefore, no method factor for the respective globally phrased scores was specified. Method factors for domain-specifically phrased scores were allowed to correlate. In Model 2, the same two factors were merged (). The models were compared using the difference in CFI (Meade, Johnson, & Braddy, Citation2008). If Model 2 fitted better, this would mean that the additional variance due to adding ‘in math’ is not instrument-specific. In other words, it would not be facet variance of achievement motivation or goal-orientation but the very same variance in each of the scales which makes up the added variance due to the phrase ‘in math’.

Figure 1. CTC(M-1) Model 1. HS: hope for success; FF: fear of failure; MAP: mastery-approach; MAV: mastery-avoidance; PAP: performance-approach; PAV: performance-avoidance; g: general; d: domain. Residuals are not shown.

Figure 1. CTC(M-1) Model 1. HS: hope for success; FF: fear of failure; MAP: mastery-approach; MAV: mastery-avoidance; PAP: performance-approach; PAV: performance-avoidance; g: general; d: domain. Residuals are not shown.

Figure 2. CTC(M-1) Model 2. HS: hope for success; FF: fear of failure; MAP: mastery-approach; MAV: mastery-avoidance; PAP: performance-approach; PAV: performance-avoidance; g: general; d: domain. Residuals are not shown.

Figure 2. CTC(M-1) Model 2. HS: hope for success; FF: fear of failure; MAP: mastery-approach; MAV: mastery-avoidance; PAP: performance-approach; PAV: performance-avoidance; g: general; d: domain. Residuals are not shown.

Consistency and method-specificity were estimated following Nussbeck, Eid, Geiser, Courvoisier, and Lischetzke (Citation2009). Accordingly, consistency can be understood as the proportion of trait variance relative to trait and method factor variance. Method specificity is the proportion of method factor variance relative to trait and method factor variance. Confidence intervals for all paths were estimated (95% intervals) in order to compare the loadings for generally and domain-specifically phrased scales.

In a second analytical step, different school grades (math, physics, German) were integrated into the preferred model (Models 2a–c) and regressed on the latent trait and method factor(s). For the math grade, this model informs us about the specific role each variance source has for the test-criterion correlation. As for the physics and German grade, those models will show whether this specific role is related to the domain itself or can be generalised across domains of scholastic performance.

Results of Study 1

CTC(M-1): model fit, consistency, method specificity, and convergent validity

Model 1 with three instrument specific method factors yielded an acceptable fit (&Chi;2[100] = 324.45, p < .001, RMSEA = 0.083, CFI = 0.937, SRMR = 0.034). Model 2 in which the method factors were merged for the domain-specifically phrased scores had a very similar fit (&Chi;2[101] = 325.81, p < .001, RMSEA = 0.083, CFI = 0.937, SRMR = 0.034). Thus, assuming the same latent source behind the different instruments did not deteriorate model fit. Consistencies and specificities from this model can be found in . It has to be noted here that the specificities for all globally phrased scores were zero. Thus, the method variable had no influence and was therefore not modelled in subsequent analyses.

Table 1. Standardised regression weights of the latent variables on the different scales, and consistency and method-specificity of these scales – Study 1.

Test-criterion correlations: role of trait and method factors

shows the results of regressing grades onto the latent trait variables and the method factor. It can be seen that fear of failure was related to the math grade, hope for success to the German grade (see Steinmayr & Spinath, Citation2009 for similar results), and the mastery approach to the physics grade. While the size of the regression weights did not differ much across domains, the standard errors and thus, the significances varied. Consequently, any conclusions regarding the importance of specific traits for specific domains should be postponed until the complete patterns have been replicated. An interesting aspect here could be the difference between math and physics on the one side and German, as the native language, on the other side.

Table 2. Regressions of grades on constructs and math-specific method factor – Study 1.

Of specific interest to the current study was the strong regression weight of the method factor for the math grade. The variance component due to the phrase ‘in math’ was strongly related to math performance but not the other grades.

Discussion of Study 1

In Study 1, we aimed at decomposing the variance of scores from generally and domain-specifically phrased achievement motivation measures. Using CTC(M-1) models, the variance was decomposed into three parts. The first part reflects interindividual trait differences. The second part contains the variance due to contextualising items. Finally, unsystematic measurement error was also modelled. The results show that the variance added by including the phrase ‘in math’ is not instrument-specific (research question 1). The model in which only one latent variable was used fitted equally well. Moreover, the influence of the method on generally phrased items was negligible. Second, the findings attest to strong trait loadings and convincing convergent validity between the corresponding domain and generally phrased scores (research question 2). However, the models in which grades were predicted imply that the incremental validity of domain-specifically phrased scale scores comes from the variance introduced by ‘in math’ (research question 3). The findings further show that adding this phrase had more influence on positively connoted traits (hope for success and mastery approach) than negatively connoted traits. The results also replicate the stronger test-criterion correlations with the math grade for domain-specifically phrased scales on the manifest level. Within the CTC(M-1) models, it could be seen that especially the variance captured in the method factor is related to variance in math grades (research question 4). This specific relation does not generalise across grades.

Study 2 will follow up on these findings and test how the method variance due to item contextualisation relates to other constructs.

Limitations

Despite the relative breadth in constructs, we operationalised each construct with one measure only. This was done to keep the total number of items at a minimum. However, the results show that consistency and specificity varied. Thus, different measures from the ones used here might also yield different relative variance proportions. Second, we only operationalised domain-specificity with regard to math. Whether the findings generalise to contextualising items for other school subjects remains an open question, which we set out to tackle in Study 2. Finally, the sample consisted of rather young students from heterogeneous school types. While we tried to control for the latter statistically, both aspects might have led to variance inflations which would lead to overestimations. Consequently, Study 2 utilised a more homogeneous sample with regard to age and school type.

To sum up, Study 1 revealed that domain-specifically phrased instruments include substantial amounts of variance which can be interpreted as results of contextualising items by including the phrase ‘in math’. It could be shown that this variance component is not instrument-specific. However, it was related to scholastic performance. The nature of this variance, modelled as a method factor here, remains unclear, even though positively connoted traits seem to be more affected by it.

Study 2

The aims of Study 2 were to replicate the findings of Study 1 using a sample of older and with regard to school type more homogeneous students. At the same time, items were contextualised not only for math but also for German. Moreover, based on theoretical grounds, we explored possible correlates of the method factor.

Possible correlates of the method factor

Study 1 revealed that the variance captured in the method factor is related to grades in the respective school subject, at least for math. Thus, when it comes to gauging the nature of this method factor, two alternative hypotheses seem viable. First, it could be assumed that the deviation from the generally phrased scales reflects a person’s standing on a lower order facet of the trait. Alternatively, it could be assumed that by contextualising items, additional constructs are being tapped. This raises the question as to which constructs might be tapped when adding phrases such as ‘in math’.

We selected potential additional constructs based on two aspects. First, as McClelland et al. (Citation1989) already proposed, self-reported achievement motivation might be nothing but self-concept. With regard to domain-specific achievement motivation, the additional method variance might tap into domain-specific self-concepts. Research on self-concept shows that it is organised in a domain-specific way, relevant to the prediction of scholastic performance, and related to achievement motivation (Helmke & van Aken, Citation1995; Jansen, Schroeders, & Lüdtke, Citation2014, Marsh & O'Mara, Citation2008; Suárez-Álvarez, Fernández-Alonso, & Muñiz, Citation2014). An interesting observation is that most self-concept scales achieve domain-specificity by contextualising items in the same way as it is done for achievement motivation scales, i.e. by adding phrases such as ‘in math (class)’ (e.g. the widely used Academic Self-Description Questionnaire II by Marsh, Citation1990b).

Another similarity between self-concept scales and the method factor specified here is that both overlap with differences in grades. Earlier findings suggest that not only self-concept, but also self-esteem is related to grades., indicating that self-concept and self-esteem are distinguishable constructs (Jansen, Scherer, & Schroeders, Citation2015; Scherer, Citation2013). For investigating whether contextualising items by adding ‘in math’ introduces variance from additional constructs, both domain-specific self-concept and self-esteem are promising candidates. As several reviews underline the difficulties to distinguish self-concept and self-esteem, both were included here (Byrne, Citation1996; Valentine, DuBois, & Cooper, Citation2004). Finally, both constructs are positively connoted which fits the results of Study 1.

Analytical strategy

Ziegler and Hagemann (Citation2015) highlighted the importance of testing the discriminant validity of test scores by including target and supposedly discriminant constructs within the same structural model. Ziegler and Bäckström (Citation2016) also suggested following this approach when looking at test-criterion correlations of facets. This way, specific relations controlled for unwanted construct overlap can be identified. Based on the explanations above and these suggestions, it seems promising to test whether the variance added through contextualising achievement motivation scales can be discriminated from self-concept and self-esteem when looking at test-criterion correlations. Such a head-to-head comparison should leave sufficient specific variance to maintain a correlation between the method factor identified in Study 1 and scholastic performance criteria, if the method factor reflects something different from self-concept and self-esteem.

Prior research on the interplay of motivational variables as predictors of specific school grades

Research looking at domain-specific measures of motivational constructs has a long and rich tradition. While a lot of work directly focuses on test-criterion correlations (e.g. Bong, Citation2002a, Citation2002b; Lee, Lee, & Bong, Citation2014; Pajares & Miller, Citation1994, Citation1995; Steinmayr & Spinath, Citation2007, Citation2009) other research has focussed on more psychometric aspects. For example, temporal stability (Bong, Citation2005) or domain-specificity (Bong, Citation2001, Citation2002a, Citation2002b; Green et al., Citation2007) have been scrutinised. Those studies show that there is temporal stability and strong convergent validity evidence for different domain-specific scales. Other studies have looked at common antecedents of domain-specific motivational constructs (Trautwein, Lüdtke, Marsh, Köller, & Baumert, Citation2006) and identified domain-specific motivational styles (Galloway, Leo, Rogers, & Armstrong, Citation1996) or epistemological beliefs (Buehl & Alexander, Citation2005). There is also research investigating gender-specific differences in test-criterion correlations (Spinath, Eckert, & Steinmayr, Citation2014; Spinath, Freudenthaler, & Neubauer, Citation2010). Valentine et al. (Citation2004) reviewed results on the relation between self-beliefs, including self-concept, self-esteem, and self-efficacy, and achievement on general and domain-specific levels. There was a small but significant regression coefficient (β = 0.08) between self-beliefs and academic achievement on general level when controlling for prior achievement. The relation grew to a medium effect size for domain-specific measures. Seaton, Parker, Marsh, Craven, and Yeung (Citation2014) investigated the interplay of math specific self-concept and math specific achievement motivation as predictors of math performance. Their findings were that math specific self-concept and math specific achievement motivation were moderately correlated as well as that math specific achievement motivation was not predictive of math performance over and above math specific self-concept. This is of relevance for Study 2. A replication of Seaton et al.’s findings would further support the hypothesis that adding ‘in math’ also adds another construct to the measures of achievement motivation, rather than capturing a facet level. However, despite all of this important and fruitful work, a systematic crossover, comparing domain-specifically phrased with general measures using modern variance decomposition methods to take a closer look at the source of test-criterion correlations, is missing.

Aims of Study 2

The first aim of Study 2 was to replicate results from Study 1 and overcome some of its limitations. The second aim of Study 2 was to put domain-specific achievement motivation, self-concept, and self-esteem in a head-to-head comparison regarding test-criterion correlations. To this end, we again applied a CTC(M-1) approach to test whether the method variance due to adding phrases like ‘in math’ would still predict scholastic performance after controlling for self-concept and self-esteem.

Method of Study 2

Sample and power

The sample consisted of 390 students attending a German secondary school type that prepares for university (Gymnasium) (231 girls, average age 17years [SD = 1.06]). The sample was recruited at two schools in two midsized towns, each testing two subsequent cohorts. Students participated voluntarily and their parents signed consent forms.

Again, a-priori power analyses considered the CTC(M-1) models and the regression. For the CTC(M-1) model a sample size of N=402 was estimated to have sufficient power for a model with 5 degrees of freedom. The current sample size is just below the maximum of N=402 needed.

Measures

Descriptive statistics and internal consistencies can be found in Appendix 4.

Achievement motivation

A German version (Dahme, Jungnickel, & Rathje, Citation1993) of the Achievement Motives Scale (AMS) by Nygård and Gjesme (Citation1973) was applied to measure general and specific achievement motivation via self-report. See Appendix 3 for more information about the scale and its adaptation.

Academic self-concept

Academic self-concept was assessed by the Scales for the Assessment of Academic Self-Concept (SESSKO; Schöne, Dickhäuser, Spinath, & Stiensmeier-Pelster, Citation2003) in a domain-specific version (Steinmayr & Spinath, Citation2007, Citation2008, Citation2009, see Appendix 3 for further information).

Scholastic performance

As in Study 1, self-reported grades (from 1 to 6) in math and German were used as criteria for scholastic performance. As before, grades were inverted to facilitate interpretation.

Statistical analyses

As in Study 1, a robust maximum-likelihood estimator (MLR) was used. Missing values were dealt with by using the full information maximum likelihood method (FIML). The CTC(M-1) model tested contained two correlated latent trait variables, i.e. fear of failure and hope for success. The general scale version was chosen as the reference method. Thus, two method factors were specified, one representing deviations from the reference method when contextualising items for math and one representing deviations when contextualising items for German. Based on this model we estimated consistency and method-specificity. In the next step, we included either the math or the German grade and used it in a hierarchical latent regression. In the first block, we entered both latent traits and the domain-specific latent method factor as predictor variables. In a second block, we entered the respective domain-specific self-esteem and self-concept scores as additional predictor variables. Model fit was evaluated as in Study 1.

Results of Study 2

shows the correlations between all variables. The main pattern for hope for success was that the higher a person scored on this subscale, the better the grade in this domain. For fear of failure, this pattern was reversed. As before, the correlations were higher when grade and scale score matched with regard to the domain. With regard to self-esteem, the correlations between different domain scores were comparable to prior research (Bong, Citation2002a, Citation2002b; Bong & Hocevar, Citation2002) as were the correlations to criteria (Pajares & Miller, Citation1994, Citation1995).

Table 3. Bivariate correlations between all specific predictors and zero-order correlations between all specific predictors and the criteria – Study 2.

The specified CTC(M-1) model fit the data well: &Chi;2[5] = 8.82, p < .117, RMSEA = 0.044, CFI = 0.992, SRMR = 0.028. Consistency was generally high and method-specificity low (). Nevertheless, there were substantial loadings from the latent method factor for each domain. shows standardised regression weights for each block in both hierarchical latent regressions. It can be seen that both trait variables, as well as method factors, were related to the respective grade. All in all, the amount of explained variance for math grades was larger, mainly due to the regression weight of the method factor.

Table 4. Standardised regression weights of the latent variables on the different scales, and consistency and method-specificity of these scales – Study 2.

Table 5. Regression results – Study 2.

When adding domain-specific self-concept and self-esteem scores, the amount of explained variance for the grades increased strongly for each domain. Both domain-specific self-concept and self-esteem had substantial and significant regression weights in all domains. In addition, with self-concept and self-esteem scores in the models, the relation between the grades and the respective method factors was strongly reduced and no longer significant.

Discussion of Study 2

Study 2 was conducted in order to replicate findings from Study 1 using a more homogeneous sample with regard to age and school type as well as an additional contextualisation (‘in German’). The CTC(M-1) model again yielded convergent validity evidence. The amounts of consistency were roughly the same. The method factor variance again had substantial relations with grades. Most importantly, this relation was strongly reduced when domain-specific self-concept and self-esteem scores were controlled for.

Correlates of the method factors

The current results show that the role of each domain-specific method factor for the test-criterion correlation strongly deteriorates when controlling for domain-specific self-concept and self-esteem. Moreover, the scores for those constructs not only have strong regression weights, their inclusion substantially improves the regression model. Based on the procedural suggestions by Ziegler and Bäckström (Citation2016) as well as Siegling, Petrides, and Martskvishvili (Citation2015), this implies that the variable shown to be inferior in a head-to-head comparison might be redundant. Study 2 also confirms earlier findings (Seaton et al., Citation2014).

At least two alternative explanations for the findings and thus, arguments against the idea that domain-specific achievement motivation items capture domain-specific self-concept and self-esteem could be brought forward. First of all, it could be argued that it is actually the other way around, i.e. domain-specific measures of self-concept and self-esteem actually capture achievement motivation. Considering the well-replicated findings regarding the structure of self-concept this seems unlikely, though. The idea is further contradicted by the fact that domain-specific self-concept and self-esteem both retained most of their test-criterion correlations in the regression analyses while the method factor, allegedly domain-specific achievement motivation, did not. Second, it could be argued that the present findings simply show that domain-specific achievement motivation, self-concept, and self-esteem items share method variance due to the fact that all measures use the same phrase such as ‘in math’. However, findings from Pomerance and Converse (Citation2014) can be used to refute such an alternative explanation. In contrast to our findings in the domain of motivation, they found that school-specific extraversion and conscientiousness scores kept their test-criterion correlations when domain-specific self-concept scores were added as additional predictors of scholastic performance, leadership, and health. The fact that the contextualised personality trait scores kept their test-criterion correlations makes the assumption unlikely that method variance due to adding ‘in math’ causes an overlap which decreases incremental validity.

Limitations

We only used two different operationalisations of achievement motivation, as the data were collected with several research questions in mind, thereby requiring a compromise regarding the total number of instruments. As a consequence, we could not replicate the finding that the method factor was not instrument-specific. Thus, future research needs to aim at replicating this. Moreover, self-concept and especially self-esteem were assessed with very few items. Thus, their role might be underestimated. However, recent research on test-criterion correlations of short scales supports the test-criterion related validity of even very short scales (Thalmayer, Saucier, & Eigenhuis, Citation2011; Ziegler, Poropat, & Mell, Citation2014).

One potential limitation of both studies not discussed so far is the use of self-reported grades instead of achievement test results. Such grades are potentially less objective. However, several studies showed that self-reported grades provide a valid representation of students’ actual performance (Dickhäuser & Plenter, Citation2005; Kuncel, Credé, & Thomas, Citation2005).

General discussion

The current research project aimed at adding to the literature on domain-specific achievement motivation and goal-orientation scales. Prior research had shown that contextualising such instruments yields improved test-criterion correlations. Before our studies, there was little empirical evidence supporting the claim that by contextualising items, specific facets of those traits are being tapped. In order to overcome this research gap, we utilised CTC(M-1) models in order to decompose scale variance and correlate the different variance components with grades. The results show that adding phrases such as ‘in math’ adds systematic variance which is related to grades and that this relation strongly deteriorates when controlling for self-esteem and self-concept in the respective domain. Moreover, the results imply that the additional variance is not instrument- or construct specific. Together these findings cast doubts on the interpretation of the additional variance component as reflecting domain-specific trait facets of achievement motivation.

Thus, the open question is, what exactly is being added by contextualising achievement motivation and goal-orientation items? Using models of the item response process we generate a hypothesis for future research: After a mental representation of the item content is built, it is compared to information retrieved from memory and a general judgment is made and mapped onto the response scale (Krosnick, Citation1999; Ziegler, Citation2011). According to the present findings, we hypothesise that the mental representation of a domain-specific achievement motivation or goal-orientation item including a phrase like ‘in math’ might evoke the retrieval of information reflecting a general achievement motive or general goal-orientation and, in addition, information on domain-specific self-concept and self-esteem. Taking the general level as a starting point, the judgment could then be fine-tuned by domain-specific self-concept and self-esteem. This hypothesis would be in line with the idea of a CTC(M-1) model which specifies the added variance as the deviation from the scores expected based on the general trait. There is also evidence from other studies supporting this hypothesis. Duda and Nicholls (Citation1992) showed that while beliefs regarding the necessity of interest, cooperation, and effort as underlying fundamentals of achievement motivation are comparable for sports and academic motivation, ability beliefs differ by domain. This would reflect the hypothesised answer process in which general motivation beliefs are coupled with domain-specific self-concept and self-efficacy.

Of course, our hypothesis does not necessarily mean that there is no such thing as a domain-specific achievement motivation or goal-orientation facet. In fact, one could argue that the results might be different for constructs not operationalised here. Still, the current research implies that using phrases such as ‘in math’ contextualises items, and that this contextualisation might not reflect differences on facet level for the constructs operationalised here. Based on our results, it seems more likely that additional variance due to different constructs like domain-specific self-esteem and self-concept is tapped. The fact that these findings might be confined to the constructs of achievement motivation and goal-orientation is highlighted by prior research on contextualising personality items which did not yield comparable results.

It has to be kept in mind that we only present limited empirical support for these assumptions. Thus, replications and thorough tests of our hypothesis are needed. Until our findings are replicated, it might be advisable to include motivational as well as self-concept and self-efficacy measures into a test battery to fully exploit the potentially predictive variance and also to find specific effects.

Conclusion

Prior research showed that domain-specifically phrased achievement motive scale scores yield better test-criterion correlations when predicting scholastic performance. Our studies built on this research and aimed to contribute towards a clearer picture of the specific contributions to test-criterion correlations different variance components in those test scores might have. Our results suggest that the higher test-criterion correlations of the domain-specific instruments could be related to additional, domain-specific variance captured. However, our findings also imply that this variance strongly overlaps with interindividual differences in domain-specific self-concept and self-esteem. If these findings were replicated, the idea of domain-specific achievement motivation facets would have to be seriously scrutinised.

Disclosure statement

No potential conflict of interest was reported by the authors.

Additional information

Funding

We acknowledge support by the Open Access Publication Fund of Humboldt Universität zu Berlin.

References

 

Appendix 1. Descriptive statistics, internal consistencies, and correlational patterns of general and domain-specific achievement motivation and grades.

Appendix 2.

Adapting scales to domain-specificity in Study 1

Adaptation of the Achievement Goal Questionnaire (Elliot & McGregor, Citation2001). The four scales have three items each with a response format ranging from 1=strongly disagree to 4=strongly agree. We chose to use an even number of answer alternatives to avoid critical issues related to the midpoint of a rating scale (Kulas & Stachowski, Citation2009; Rost, Carstensen, & von Davier, Citation1999). To construct math-specific items, the phrase ‘in math class’ was added (e.g. ‘My goal in math class is to get a better grade than the other students’.).

Adaptation of the Revised German Achievement Motives Scale (AMS-R; Lang & Fries, Citation2006; cf. Nygård & Gjesme, Citation1973). The revised German Achievement Motives Scale ( AMS-R; Lang & Fries, Citation2006; cf. Nygård & Gjesme, Citation1973) measures the motives hope for success (HS) and fear of failure (FF) with 10 items that are answered on a 4-point scale ranging from 1=strongly disagree to 4=strongly agree. Math-specific items were constructed by adding ‘in math’ (e.g. ‘In math, I like it when I can find out how capable I am’).

Appendix 3.

Adapting scales to domain-specificity in Study 2

Achievement motivation: German version (Dahme et al., Citation1993) of the Achievement Motives Scale (AMS) by Nygård and Gjesme (Citation1973). The original instrument uses 14 items each to assess hope for success and fear of failure (McClelland et al., Citation1953). In this study a short form was applied, measuring each subscale by seven items on a 4-point Likert-type scale ranging from 1=strongly disagree at all to 4=strongly agree. The factor structure of the German short version has been confirmed and the questionnaire has already been applied in other empirical studies (Steinmayr & Spinath, Citation2008, Citation2009). The original items are worded in a general fashion. In order to assess domain-specific achievement motivation the original items were combined with three answer sections titled by ‘In general’, ‘mathematics’, and ‘German’. It is worth noting that this implies that the process of contextualising items was slightly different to Study 1. There, the phrase ‘in math’ was added to each of the items.

Academic self-concept: Scales for the assessment of academic self-concept (SESSKO; Dickhäuser, Schöne, Spinath, & Stiensmeier-Pelster, Citation2002; Schöne et al., Citation2003). The authors of the SESSKO scales defined academic self-concept as ‘all cognitive representations of one’s own abilities in academic achievement situations’, for example at school (Dickhäuser et al., Citation2002, p. 394). The subscale called self-concept absolute was used in Study 2. Participants rated on a five-point rating-scale how good they think they are at school in general, in mathematics and in German, respectively. General and domain-specific self-concept in math and German were assessed by four items each. The general items were taken from the described questionnaire (Dickhäuser et al., Citation2002; Schöne et al., Citation2003).

Domain-specific self-esteem: The item ‘I have high self-esteem’ was combined with three answer sections titled by ‘In general’, ‘mathematics’, and ‘German’, to construct domain-specific self-esteem items. Each answer section had a 4-point Likert-type scale ranging from 1=strongly disagree to 4=strongly agree.

 

Appendix 4. Ranges, means, standard deviations, and internal consistencies for all used variables in Study 2