145
Views
2
CrossRef citations to date
0
Altmetric
Original Research

Validity of a screening tool for detecting subtle cognitive impairment in the middle-aged and elderly

, , &
Pages 2165-2176 | Published online: 12 Dec 2014

Abstract

The present study tested 121 middle-aged and elderly community-dwelling individuals on the computer-based Subtle Cognitive Impairment Test (SCIT) and compared their performance with that on several neuropsychological tests. The SCIT had excellent internal consistency, as demonstrated by a high split-half reliability measure (0.88–0.93). Performance on the SCIT was unaffected by the confounding factors of sex, education level, and mood state. Many participants demonstrated impaired performance on one or more of the neuropsychological tests (Controlled Oral Word Association Task, Rey Auditory and Verbal Learning Task, Grooved Pegboard [GP], Complex Figures). Performance on SCIT subtests correlated significantly with performance on many of the neuropsychological subtests, and the best and worst performing quartiles on the SCIT subtest discriminated between good and poor performers on other subtests, collectively indicating concurrent validity of the SCIT. Principal components analysis indicated that SCIT performance does not cluster with performance on most of the other cognitive tests, and instead is associated with decision-making efficacy, and processing speed and efficiency. Thus, the SCIT is responsive to the processes that underpin multiple cognitive domains, rather than being specific for a single domain. Since the SCIT is quick and easy to administer, and is well tolerated by the elderly, it may have utility as a screening tool for detecting cognitive impairment in middle-aged and elderly populations.

Introduction

There is an increasing requirement to be able to assess cognitive function in the elderly. Reasons range from detecting early stages of dementia to testing for the adverse effects of medications and major surgery.Citation1,Citation2 A survey of geriatric specialists found that cognitive assessments typically rely on screening tools that can be biased by language, culture, and education.Citation3 A review of cognitive screening in primary care and geriatric services in the UK and Canada concluded that better screening tools are urgently needed.Citation4

An ideal screening tool is brief in application, requires simple responses from the patient, is psychometrically robust, and is sensitive to changes across a wide range of cognitive domains.Citation2,Citation5 The search for better screening tools has favored the development of computer-based tests,Citation6 as they can provide uniformity of administration, accurate recording of responses, and objective scoring.Citation7Citation9 Several computerized test batteries (eg, MicroCog, CogState, CANTAB) are now widely used to detect cognitive impairment.Citation10,Citation11

Wild et al drew attention to the need to establish the validity and reliability of computer-based tests of cognition that are intended for use in elderly populations.Citation9 They reported that while many computer-based tests had demonstrated test validity, other measures of quality were not well represented. For example, normative data were inadequate for just over half of the test batteries reviewed (due to small sample sizes or lack of data specific to older adults in a larger sample), and reliability was usually only demonstrated in one form. CANTAB,Citation12 CogState,Citation13 and MicroCogCitation14 were among the small number of computer-based tests that were rated highly by Wild et al.

The Subtle Cognitive Impairment Test (SCIT; NeuroTest.com)Citation15 is a brief, computerized, visual discrimination task.Citation16 It was originally developed as a means of detecting cognitive impairments that are too slight to qualify as mild cognitive impairment (MCI), and which may be present up to 15 years before the deficits associated with MCI can be detected. Subtle cognitive impairment has been referred to as “subjective-cognitive impairment”.Citation17 These subtle cognitive impairments are objectively identifiable impairments in cognitive performance in individuals whose score on the Mini-Mental State Examination (MMSE) falls within the range that is generally taken to represent “normal” cognitive function in older persons (scores of 29–25).Citation18 The SCIT can be administered by untrained personnel, successful completion requires no previous knowledge of computers, and testing can be completed within 3–4 minutes. When a visual stimulus is presented on the computer screen, the participant decides which line is shorter and presses the corresponding left or right button.

The SCIT has been employed with a range of populations, including the elderly, children with developmental disorders, human immunodeficiency virus-1 immunopositive individuals, cardiac surgery patients, and individuals who have been sleep-deprived or are intoxicated.Citation16,Citation18Citation21 While the primary advantage of the SCIT is its rapid administration time (3 minutes compared with 15–120 minutes for other computerized measures of global cognitive function), other advantages include a lack of cultural or sex bias and lack of a learning effect that enables the SCIT to be used repeatedly without any loss of reliability.

High test-retest reliability has already been established for both the SCIT response time (0.98) and error rate (0.91) measures,Citation18 and performance on SCIT has shown medium correlations against performance on subtests of the CANTAB (eg, simple reaction time, r[57]=0.46, P<0.01; choice reaction time, r[57]=0.54, P<0.01).Citation16 However, performance on the SCIT has not been systematically compared against other neuropsychological tests that are used in research and clinical practice. The present study examines a sample of community-dwelling individuals ranging from middle-aged to old-aged and, for the purposes of assessing validity, compares their performance on the SCIT with that on several neuropsychological tests. Although the participants were community-dwelling, a considerable degree of individual variability was observed in their performance on the neuropsychological tests, and this heterogeneity provided a sufficient range of cognitive function to compare performance on those tests with that on the SCIT. The SCIT has been shown to be particularly suited to the detection of slight decrements in cognitive performance within cognitively “normal” elderly, and is sensitive to impairments in several cognitive domains, including attention, visuospatial processing, and language.Citation18

This study provides an assessment of split-half reliabilities for the SCIT, and assesses two measures of concurrent validity, as well as construct validity using confirmatory factor analysis. The results indicate that the SCIT meets all of the requirements described by Wild et al for a computer-based test of cognition that is suitable for use in elderly populations.Citation9

Materials and methods

Participants

The 121 participants in this study (76 males, 43 females) were aged 40–85 (mean ± standard deviation [SD] 64±9.1) years, had received an average of 11±2.9 (range 5–20) years of formal education, and were fluent speakers and readers of English (). All participants lived independently in the community and were recruited and assessed in Melbourne, Australia. Individuals were excluded from participating if they had a history of a neuropsychological, psychiatric, or neurological disorder, a head injury, or cardiac surgery. Participation was voluntary and all participants gave informed consent in accordance with National Health and Medical Research Council ethical guidelines.

Figure 1 Distribution of age (A) and education levels (B).

Figure 1 Distribution of age (A) and education levels (B).

Assessment tasks

The neuropsychological tests included in the test battery were chosen by an independent clinical neuropsychologist (Dr Greg Savage), on the basis of their brief duration and widespread use in the assessment of cognitive performance in relatively high-functioning individuals.

SCIT

Participants are asked to correctly identify which of two lines presented on a computer screen is shortest (). The stimulus is repeatedly presented in a pseudo-random order for exposure durations in the range of 16–176 msec, in 16 msec increments. The entire testing session lasts 3–4 minutes. Two sets of data are obtained, ie, the number of errors made at each stimulus exposure time (% error), and the time taken to respond at each stimulus exposure time (response time).

Figure 2 An example of the sequence of presentation of the Subtle Cognitive Impairment Test stimuli (focal point, blank screen, one version of the test stimuli, and backward mask) and durations of presentation of each. In this example, the short arm of the test stimulus is on the left.

Figure 2 An example of the sequence of presentation of the Subtle Cognitive Impairment Test stimuli (focal point, blank screen, one version of the test stimuli, and backward mask) and durations of presentation of each. In this example, the short arm of the test stimulus is on the left.

The first four exposure durations (16–64 msec) are referred to as the “head” of the data curve. Data from these four exposure durations (16–64 msec) are combined to provide two representative subtest scores, ie, the error rate in the head of the data curve (SCIT-EH) and response times in the head of the data curve (SCIT-RTH).Citation18,Citation19 The remaining seven exposure durations (80–176 msec) represent the “tail” of the data curve and are pooled to provide two further representative subtest scores, ie, error rates in the tail of the data curve (SCIT-ET) and response times in the tail of the data curve (SCIT-RTT).

Wechsler Test of Adult Reading

The Wechsler Test of Adult Reading (WTAR) is a 50-word list designed to estimate premorbid intellectual function. Participants are asked to read 50 words aloud and are assessed on their correct pronunciation. The task is scored by summation of all correct responses, where correct responses are scored as 1 and incorrect responses are scored as 0 (reliability, split-half =0.93; test-retest =0.94).Citation22

Depression, Anxiety and Stress Scale

The Depression, Anxiety and Stress Scale (DASS) is a 21-item questionnaire that assesses the negative emotional states of depression, anxiety, and stress. This test was used because each of these emotional states may affect cognitive performance. Participants are given a questionnaire with 21 statements relating to their emotional state during the previous week. The statements equally represent depression (DASS-D), anxiety (DASS-A), and stress (DASS-S). The participant indicates how often in the past week each statement applied to them. A “0” indicates the statement does not apply to them, “1” is for some of the previous week, “2” is a good part of the previous week, and “3” is for most of the previous week. Each emotional state can score a maximum of 21 points (reliability [Cronbach’s alpha], depression =0.91, anxiety =0.84, stress =0.90).Citation23

Controlled Oral Word Association Task

The Controlled Oral Word Association Task (COWAT) is a measure of phonemic verbal fluency.Citation24,Citation25 Participants are presented with a letter of the alphabet and have 1 minute to produce as many words as they can that begin with the letter. This process is performed with three different letters, with a score being derived from the total number of correct words produced (reliability, internal consistency =0.83; test-retest =0.74).Citation26

Grooved Pegboard

The Grooved Pegboard (GP) measures motor coordination and dexterity.Citation27 Participants are required to use their dominant hand (GP-D) to correctly insert pegs into a pegboard in a certain sequence or pattern. The time taken is recorded, and then the process is repeated with their nondominant hand (GP-ND). The total score for this task is calculated as the number of seconds taken to complete the task plus the number of pegs dropped plus the number of pegs correctly placed (reliability, test-retest =0.82).Citation28

Medical College of Georgia Complex Figures

The Medical College of Georgia Complex Figures (MCGCF) assesses visuospatial memory and perceptual organization.Citation29Citation31 Participants are required to copy a picture and to remember as much of it as they can (MCG-C; Copy Trial). The picture is removed and they are then asked to draw as much of the picture as they can freely recall (MCG-I; Immediate Recall Trial). Participants are presented with other tasks and measures, then after a 30-minute delay they are asked to recall and redraw the picture (MCG-D, Delayed Recall Trial). The picture consists of 18 elements, each of which can be scored from 0 (not recalled) to 2 (correctly placed and correctly drawn) for a maximum score of 36 points per trial. (Reliability, test-retest; copy =0.32; immediate recall =0.71; delayed recall =0.73).Citation32

Rey Auditory and Verbal Learning Task

The Rey Auditory and Verbal Learning Task (RAVLT)Citation33,Citation34 assesses overall memory performance, immediate memory span, acquisition rate, interference effects, and recognition memory.Citation25,Citation31 Participants are read a list of 15 words (List A) and are asked to freely recall as many as they remember. This process is performed five times and provides the first set of results that are identified as Learning Trials I–V (RAVLT-L). Participants are then presented with a second list of 15 words (List B) and subsequently asked to freely recall List A words, this score is presented as Immediate Recall Trial (RAVLT-I). After a 20-minute delay, participants are asked to freely recall List A words, these results constitute the Delayed Recall Trial (RAVLT-D). Finally, participants are read a list of 50 words, containing words from List A, List B, and distracter words that are phonetically and/or semantically related to words in either List A or B. Participants are required to identify List A words only, this being the Recognition Trial (RAVLT-R). (reliability, test-retest; learning trials =0.72–0.78; immediate recall =0.67–0.81; delayed recall =0.71–0.81; recognition trial =0.38–0.66).Citation35

Procedures

The total assessment took approximately 45 minutes to perform and was conducted by the same researcher in a quiet room or office. All participants performed the tests in the following order: WTAR, MCG-C, MCG-I, RAVLT-L, RAVLT-I, SCIT, COWAT, DASS, RAVLT-D, RAVLT-R, MCG-D, GP-D, and GP-ND. The order was due to the timing constraints imposed by the delayed recall trials of the MCG and RAVLT.

Statistical analysis

All analyses used Statistical Package for the Social Sciences version 21 software (IBM Corporation, Armonk, NY, USA). Initial descriptive statistics identified the range of data obtained for each measure (). To determine the internal consistency of the SCIT, a split-half reliability coefficient was calculated and adjusted using the Spearman-Brown formula. Concurrent validity was examined by correlating participants’ performance on each of the four SCIT subtests with their performance on each of the neuropsychological subtests. In addition, a series of t-tests was used to determine whether the best and worst performers on the SCIT subtests (first quartile and fourth quartile data, respectively) corresponded with good and poor (respectively) performances on the other neuropsychological subtests. Where multiple comparisons were undertaken, type I error was controlled for with the false discovery rate test.Citation36,Citation37 A principal components analysis examined which neuropsychological subtests the SCIT subtests cluster with (convergent validity) and which neuropsychological subtests the SCIT does not cluster with (divergent validity).

Table 1 Sample size, mean, SD, and range of participant scores for the measures assessed in the present study

Results

Sample descriptive statistics

Summary data for each of the assessment measures used in the study are shown in . They include the number of participants that completed each measure, the mean score for each measure, and a variety of estimates describing the distribution of the data. The latter include the SD, range (ie, minimum and maximum scores), the highest score possible for each measure, the ratio of the SD to the mean (which allows comparison of relative variability across measures), and the skewness of the distribution.

Each of the 121 participants completed most or all of the measures (). One participant did not complete the COWAT, being uncomfortable with the task and requesting not to continue. Four participants did not complete the RAVLT, and a further three participants chose not to complete the recognition trial of the RAVLT; eleven and 13 participants were unable to complete the GP-D and GP-ND tests, respectively. Reasons for this lack of participation included: the task being unavailable when the assessment took place (n=4), the test being too difficult for the participant to complete (n=2), and participants having an impediment (an injury to hand or part of hand/fingers missing; n=3), or osteoarthritis (n=5).

For the majority of the measures, there was sufficient range and variance in the scores (both in terms of the SD and the ratio of the SD to the mean), and the skew of their distribution was within the acceptable range (−1 to +2; ). Subtests that initially appeared not to meet one or more of these criteria were the MCG-C, RAVLT-R, SCIT-ET, and the three DASS measures. The MCG-C scores had a restricted range and the distribution of scores had a negative skew in excess of −1. Since the copy trial of the MCGCF requires participants to copy a simple figure that remains directly in front of them, the majority of participants from this population were expected to complete this subtest with few errors. Thus a restricted range on the MCG-C is what is expected, and the negative skew is the consequence of a few participants having a deficit on this very simple task. The restricted variance and negative skew on the recognition trial of the RAVLT was also expected. Recognition memory generally exceeds recall memory, and consequently, participants should perform more accurately on the recognition trial and the recall trials. Given that there is a ceiling on how many responses are available, the range of scores on the recognition trials is more constrained than that on the recall trials, and the spread of scores is constrained by this ceiling, generating a negative skew.

Both the DASS subtest and SCIT-ET subtest had a wide range of scores but the distribution of the scores was positively skewed. This pattern was expected for the DASS subtest as the majority of participants were emotionally stable, but a few individuals were moderately stressed, depressed, or anxious. For the SCIT-ET subtest, it was expected that the majority of participants would make no errors. Hence data for the SCIT-ET are not expected to be normally distributed and the spread generated by the small number of participants with greater errors produces the higher skew value. In view of the foregoing considerations, it was concluded that the restricted range and/or skew of the MCG-C, RAVLT-R, DASS, and SCIT-ET data is a natural product of those measures, so transformation of the data was not required.

Split-half reliability

The internal consistency of SCIT was determined by split-half reliability. For each participant, the response time and error data for the first half of the test items (n=55) were correlated against that for the remaining (n=55) items. Importantly, while each participant received the items in a different random order, each half of the test contained an equal number of items at each of the exposure durations and an equal number of left and right stimuli. A comparison across all of the participants revealed that the SCIT has very high split-half reliabilities of r(53)=0.87 (P<0.01) for response time and r(53)=0.79 (P<0.01) for error rate. Application of the Spearman-Brown adjustment results in internal consistency reliabilities of 0.93 for response time and 0.88 for error rate.

Validity

In order to assess concurrent validity, performance on the four SCIT subtests was correlated with that on the eleven other neuropsychological subtests. After the effect of age had been removed and type I error corrected, performance on each of the SCIT subtests was found to correlate significantly with performance on many subtests from each of the neuropsychological measures included in the study (). They were all small-to-medium level correlations, indicating that there is an association between performance on the neuropsychological subtests and the SCIT subtests.

Table 2 Correlations of the four SCIT subtests with other neuropsychological measures

Performance on both of the SCIT error subtests correlated with more neuropsychological subtests than did the SCIT response time subtests. Further, performance in the tail of the SCIT curve (SCIT-RTT and SCIT-ET) correlated with more subtests than performance in the head (SCIT-RTH and SCIT-EH). Performance on all four SCIT subtests correlated positively with age and performance on the GP-D, and correlated negatively with performance on the immediate (MCG-I) and delayed (MCG-D) recall subtests of the MCGCF test. SCIT-RTH was also negatively correlated with the COWAT. SCIT-RTT also correlated positively with performance on the GP-D and negatively with the WTAR, and the delayed recall subtest of the RAVLT. SCIT-EH performance, in addition to correlating with age, the MCG-I, and the MCG-D, also correlated negatively with performance on the COWAT, RAVLT-L, RAVLT-I, and RAVLT-D, and positively with GP-ND performance and the depression subscale of the DASS-D. SCIT-ET mirrored the correlations seen for the SCIT-EH subtest except for the absence of a positive correlation with the DASS-D subtest and the addition of a negative significant correlation with the WTAR ().

The t-test comparisons revealed that individuals whose performances on the SCIT subtests were in the top (ie, first) quartile relative to those whose performances were in the bottom (ie, fourth) quartile were also significantly better or worse on a number of the neuropsychological subtests. The neuropsychological tests for which good or poor performance was significantly distinguished by the highest and lowest performance on the SCIT subtests (after correcting for type I error) were the COWAT, DASS-D, GP-D, GP-ND, MCG-C, MCG-I, MCG-D, RAVLT-L, RAVLT-I, and RAVLT-D. All the significant relationships between the SCIT subtests and the other neuropsychological tests are shown in . Participants in the first and fourth quartiles on the SCIT-EH subtest distinguished between good and poor performers on nine other measures. Equivalently categorized groups on the SCIT-ET subtest distinguished between good and poor performers on eight other subtests, with the SCIT-RTT and SCIT-RTH subtests differentiating performance on seven and three subtests, respectively.

Figure 3 Best and worst quartiles on each of the SCIT subtests compared with cognitive performance on all other subtests. Significant differences are shown with an asterisk.

Abbreviations: COWAT, Controlled Oral Word Association Task; DASS, Depression, Anxiety and Stress Scale; DASS-A, DASS-Anxiety; DASS-D, DASS-Depression; DASS-S, DASS-Stress; GP-D, Grooved Pegboard using dominant hand; GP-ND, Grooved Pegboard using nondominant hand; MCG, Medical College of Georgia; MCG-I, MCG Immediate Recall Trial; MCG-D, MCG Delayed Recall Trial; MCG-C, MCG Copy Trial; RAVLT, Rey Auditory and Verbal Learning Task; RAVLT-D, RAVLT Delayed Recall Trial; RAVLT-I, RAVLT Immediate Recall Trial; RAVLT-R, RAVLT Recognition Trial; RAVLT-L, RAVLT Learning Trials I–V; SCIT, Subtle Cognitive Impairment Test; SCIT-EH, error rates in the head of the data curve; SCIT-ET, error rates in the tail of the data curve; SCIT-RTT, response times in the tail of the data curve; SCIT-RTH, response times in the head of the data curve; WTAR, Wechsler Test of Adult Reading.
Figure 3 Best and worst quartiles on each of the SCIT subtests compared with cognitive performance on all other subtests. Significant differences are shown with an asterisk.

The significant correlations from and the significant associations between best and worst quartile subtest performance () together provided the basis for a Venn diagram (). This diagram summarizes the significant associations between the four SCIT subtests and the other neuropsychological subtests used in the test battery.

Figure 4 Venn diagram summarizing the outcomes from the predictive validity measures (the significant t-test outcomes comparing high and low performers on SCIT and each of the other tests) and the concurrent validity measures (the significant correlations between the four subtests of the SCIT and other measures). The variables located outside of the circles did not have any significant associations with performance on any of the SCIT subtests.

Abbreviations: COWAT, Controlled Oral Word Association Task; DASS, Depression, Anxiety and Stress Scale; DASS-A, DASS-Anxiety; DASS-D, DASS-Depression; DASS-S, DASS-Stress; GP-D, Grooved Pegboard using dominant hand; GP-ND, Grooved Pegboard using nondominant hand; MCG, Medical College of Georgia; MCG-I, MCG Immediate Recall Trial; MCG-D, MCG Delayed Recall Trial; MCG-C, MCG Copy Trial; RAVLT, Rey Auditory and Verbal Learning Task; RAVLT-D, RAVLT Delayed Recall Trial; RAVLT-I, RAVLT Immediate Recall Trial; RAVLT-R, RAVLT Recognition Trial; RAVLT-L, RAVLT Learning Trials I–V; SCIT, Subtle Cognitive Impairment Test; SCIT-EH, error rates in the head of the data curve; SCIT-ET, error rates in the tail of the data curve; SCIT-RTT, response times in the tail of the data curve; SCIT-RTH, response times in the head of the data curve; WTAR, Wechsler Test of Adult Reading.
Figure 4 Venn diagram summarizing the outcomes from the predictive validity measures (the significant t-test outcomes comparing high and low performers on SCIT and each of the other tests) and the concurrent validity measures (the significant correlations between the four subtests of the SCIT and other measures). The variables located outside of the circles did not have any significant associations with performance on any of the SCIT subtests.

A principal components factor analysis was conducted on the full set of 18 tests/subtests with orthogonal rotation (varimax). The Kaiser-Meyer-Olkin measure (=0.65) verified that the sample size was adequate for factor analysis. Bartlett’s test of sphericity (χ2[153]=1,161.78, P<0.001) indicates that the correlations between test performances were sufficiently large for principal components factoring. Six factors with eigenvalues over 1.00 were identified that cumulatively explained 77.23% of the variance. That is, the analysis revealed six distinct constructs (factors) among the 18 tests/subtests which, for the most part, delineated individual neurological tests from each other (). All the verbal learning and memory subtests (ie, RAVLT) loaded most strongly on construct 1, while all the mood state measures (ie, DASS) loaded strongly on construct 2. The measures loading strongly on construct 3 were concerned with decision-making efficacy (ie, GP-D, GP-ND, SCIT-ET) and had a small factor loading from the SCIT-EH subtest. Construct 4 was comprised of measures that assess processing speed and efficiency (ie, SCIT-RTH, SCIT-RTT, and SCIT-EH). Construct 5 consisted of the MCG subtests, which assess visuospatial learning and memory, while construct 6 contained measures of vocabulary, ie, the WTAR and COWAT.

Table 3 Principal components analysis indicates that the cognitive tests used in this study segregate into six major cognitive domains

Discussion

The present study validated the SCIT against a battery of neuropsychological tests in a group of middle-aged and elderly participants to assess its usefulness as a tool for detecting decrements in cognitive performance. A lack of knowledge of computers can lead to stress and noncompliance when elderly persons are asked to use computer-based tests.Citation8 In the present study, however, every participant completed the SCIT, whereas some participants refused to complete each of the other (non-computer-based) tests. The SCIT was found to have a high split-half reliability and was unaffected by a range of confounding factors, including sex, level of education, and negative emotional state. The SCIT was found to have almost no correlation with the DASS, showing that there is little confounding effect of mild depression, anxiety, or stress. Performance on the SCIT was influenced by age, as were all of the other tests in the study. When the effects of age had been removed, the SCIT exhibited small-to-medium level correlations with most of the other tests. Interestingly, the four SCIT subtests each had a characteristic pattern of correlations with other tests in the battery. The implications of these findings are discussed below.

Heterogeneity of the data

This study examined the cognitive performance of people who led independent lives and had not been diagnosed with a neuropsychological, psychiatric, or neurological disorder. However, few participants obtained perfect scores on any of the cognitive tests, and there was considerable variability in the scores, with a proportion of participants displaying moderate impairment on each of the tests (). Such heterogeneity in cognitive performance is typical of elderly populations, and is attributable to benign age-related impairment,Citation38 and to the presence of mild levels of cognitive impairment that may presage Alzheimer’s disease.Citation17 The wide range of scores obtained on the cognitive tests and on the SCIT enabled performance on these tests to be compared, and for meaningful correlations to be derived.

Reliability

It has already been established that the SCIT has a high test-retest reliability for both the response time subtest (0.98) and the error rate subtest (0.91).Citation16,Citation18 To comply with the recommendation that highly rated computer-based tests should provide more than one measure of reliability,Citation9 the internal consistency of the SCIT was determined by the calculation of split-half reliability for both the response time and error rate subtests. A split-half reliability coefficient of greater than 0.70 indicates high internal consistency.Citation39 The SCIT was found to have a split-half coefficient of 0.93 for response time and 0.88 for error rate. That is, the SCIT has high internal consistency in addition to high test-retest reliability.

Potential confounding factors

Before examining the relationship between performance on the SCIT and the other measures of cognitive performance, it was important to ensure that such relationships were not unduly influenced by confounding factors. Six potential confounding factors were examined: level of education, sex, age, and the negative emotional states of stress, anxiety, and depression. While some of the neuropsychological subtests used in the battery were affected by these factors, only age affected performance on the SCIT. Age also affected performance on all of the other neuropsychological tests used in this study.

It is not surprising that performance on each of the SCIT subtests correlated with age, given that increasing age is a nonmodifiable risk factor for cognitive and motor decline.Citation38 Numerous studies have reported that increasing age correlates with impaired performance across a range of cognitive and motor domains.Citation38,Citation40Citation42 Performance on three other computerized tests is also affected by age (MicroCog,Citation43 CogState,Citation10 CANTABCitation44).

Validity

Performance on the SCIT correlated with all but two of the neuropsychological subtests, ie, the recognition trial of the RAVLT and the copy trial of the MCG. All four subtests of the SCIT correlated with performance on the immediate and delayed recall trials of the MCGCF task and the GP-D, whereas the pattern of correlations with the other neuropsychological subtests differed between the SCIT subtests. These differences reflect the fact that the SCIT has both accuracy (% error rate) and speed (response time) components. These are further separated into an unconscious attention component (head part of SCIT curve) at stimulus presentation times of 16–64 msec and a conscious attention component (tail part of SCIT curve) with stimulus presentation times of 80–176 msec. This separation is based on findings from masked priming studies showing that stimuli presented at durations of 64 msec or less are processed without conscious awareness, but are still able to influence subsequent decision-making via automatic processes.Citation45Citation47 Since most of the neuropsychological tests used in the present study had a stronger requirement for accuracy than for speed, it was expected that more correlations would be obtained for the SCIT error rate subtests.

All of the significant correlations were of small or medium levels (between 0.205 and 0.466), indicating that performance on the SCIT subtests was not strongly related to, or solely driven by, the cognitive domains assessed by each neuropsychological subtest. That performance on the SCIT is correlated with performance on the subtests of the MCGCF and the GP is readily explained by the shared visuospatial nature of these tasks. However, the significant correlations between the SCIT subtest scores and measures of verbal fluency (ie, COWAT), knowledge of vocabulary (ie, WTAR), and verbal working and episodic memory (ie, RAVLT) do not have a straightforward explanation, since the SCIT is largely nonverbal. We speculate, for example, that the correlations between the RAVLT learning and recall trials and performance on the tail subtests of the SCIT (SCIT-RTT and SCIT-ET) may reflect the underlying role of attention. The exposure durations in the tail region of the SCIT curve are long enough for participants to be consciously aware of the stimulus, so it is likely that attention plays a role in the SCIT decision-making process at those exposure durations, as it does in the learning and recall trials of the RAVLT.Citation48,Citation49

The outcomes of the comparative performances on each of the neuropsychological subtests with good and poor performers on the SCIT subtests demonstrated the capacity of the SCIT subtests to distinguish between good and poor performance across the other cognitive domains tested in the study. The SCIT error rate subtests were more sensitive to performance outcomes on other cognitive tests than the response time subtests, possibly because these tests were not timed and did not have a high response time component (except for the GP).

A principal components analysis revealed that the SCIT primarily loads on one construct (convergent validity), processing speed, and efficiency, with the error components loading in full or in part on the construct of decision efficacy. Most of the other tests did not load strongly on these constructs (divergent validity). Impaired performance on the SCIT may reflect compromised signal processing speed (response time subtests) and reduced efficacy of signal processing and decision-making (error rate subtests). These speculations await confirmation by electroencephalography and functional imaging studies.

The SCIT does not have high concurrent validity against the other cognitive measures because the constructs that it measures (efficacy, speed, and efficiency of processing) are not the primary domains measured by the other tests. Despite this, performance on the SCIT correlates mildly to moderately with performance on most of the other tests, and the SCIT is able to discriminate between good and poor performers on those tests. These properties indicate that the SCIT measures constructs that are common to performance in a wide range of cognitive domains. This generality makes the SCIT useful for the early detection of global cognitive impairment, rather than impairments across specific cognitive domains, such as is done by CogStateCitation50 and MicroCog.Citation51

Computer-based cognitive tests can be compared on the availability of normative data, test validity and reliability, comprehensiveness, and usability.Citation9 The SCIT rates highly on these criteria, and consequently it may have utility as a screening tool for detection of a generalized subtle cognitive impairment; people who are identified with such a deficit can be referred for a detailed neuropsychological examination. The SCIT may have utility in cognitive screening of elderly populations, since it is well tolerated by the elderly, and performance on the SCIT has previously been shown to be sensitive to decrements in performance on the MMSE.Citation18

Conclusion

In a group of community-dwelling, middle-aged, and elderly individuals, the SCIT showed validity against well-established measures of visuospatial processing and memory (MCGCF), motor coordination and dexterity (GPB), premorbid IQ (WTAR), verbal fluency (COWAT), and verbal learning and memory (RAVLT). The broad range of significant associations indicates that the SCIT is not sensitive to a particular cognitive domain and instead provides a general measure of cognitive function. It should be noted that the SCIT is only suitable for use in high-functioning individuals, as people with an MMSE score of less than 24 are unable to complete the SCIT.Citation18 It remains to be determined whether elderly individuals who display impaired performance on the SCIT are more likely to develop dementia.

Acknowledgments

The Australian and New Zealand Society of Cardiac and Thoracic Surgeons supported this work. KB was supported by an Australian Postgraduate Award. The authors thank the participants for generously giving their time to be involved in this study.

Disclosure

The authors report no conflicts of interest in this work.

References

  • BeinhoffUHilbertVBittnerDGronGRiepeMWScreening for cognitive impairment: a triage for outpatient careDement Geriatr Cogn Disord20052027828516158010
  • MilmanLHHollandAKaszniakAWD’AgostinoJGarrettMRapcsakSInitial validity and reliability of the SCCAN: using tailored testing to assess adult cognition and communicationJ Speech Lang Hear Res200851496918230855
  • ShulmanKIHerrmannNBrodatyHIPA survey of brief cognitive screening instrumentsInt Psychogeriatr20061828129416466586
  • IsmailZRajjiTKShulmanKIBrief cognitive screening instruments: an updateInt J Geriatr Psychiatry20102511112019582756
  • MaruffPCollieADarbyDWeaver-CarginJMastersCCurrieJSubtle memory decline over 12 months in mild cognitive impairmentDement Geriatr Cogn Disord20041834234815316183
  • ButcherJNPerryJHahnJComputers in clinical assessment: historical developments, present status, and future challengesJ Clin Psychol20046033134514981795
  • BleibergJKaneRLReevesDLGarmoeWSHalpernEFactor analysis of computerized and traditional tests used in mild brain injury researchClin Neuropsychol20001428729411262703
  • WeberBFritzeJSchneiderBKuhnerTMaurerKBias in computerized neuropsychological assessment of depressive disorders caused by computer attitudeActa Psychiatr Scand200210512613011939962
  • WildKHowiesonDWebbeFSeelyeAKayeJStatus of computerized cognitive testing in aging: a systematic reviewAlzheimers Dement2008442843719012868
  • De JagerCASchrijnemaekersACHoneyTEBudgeMMDetection of MCI in the clinic: evaluation of the sensitivity and specificity of a computerised test battery, the Hopkins Verbal Learning Test and the MMSEAge Ageing20093845546019454402
  • ÉgerháziABereczRBartókEDegrellIAutomated neuropsychological test battery (CANTAB) in mild cognitive impairment and in Alzheimer’s diseaseProg Neuropsychopharmacol Biol Psychiatry20073174675117289240
  • SahakianBJMorrisRGEvendenJLA comparative study of visuospatial memory and learning in Alzheimer-type dementia and Parkinson’s diseaseBrain1988111Pt 36957183382917
  • MakdissiMCollieAMaruffPComputerised cognitive assessment of concussed Australian Rules footballersBr J Sports Med20013535436011579074
  • The Psychological CorporationMicrocog: Assessment of Cognitive Functioning (computer program). Version 2.1San Antonio, TX, USAThe Psychological Corporation1993
  • YellandGYRobinsonSRFriedmanTHutchisonCWInventorsAn automated method for measuring cognitive impairment US Patent AU20042036792004
  • FriedmanTRobinsonSRYellandGYImpaired perceptual judgement at low blood alcohol concentrationsAlcohol20114571171821145695
  • ReisbergBPrichepLMosconiLThe pre-mild cognitive impairment, subjective cognitive impairment stage of Alzheimer’s diseaseAlzheimers Dement20084S98S10818632010
  • FriedmanTYellandGRobinsonSRSubtle cognitive impairment in elders with MMSE scores within the ‘normal’ rangeInt J Geriatr Psychiatry20122746347121626569
  • SpeirsSRobinsonSRinehartNTongeBYellandGEfficacy of cognitive processes in young people with high-functioning autism and Asperger’s disorder using a novel taskJournal of Autism and Developmental Disorders2014442809281924838123
  • BruceKMYellandGWSmithJARobinsonSRRecovery of cognitive function after coronary artery bypass graft operationsAnn Thorac Surg2013951306131423333061
  • BruceKMYellandGWAlmeidaASmithJARobinsonSREffects on cognition of conventional and robotically assisted cardiac valve operationsAnn Thorac Surg201497485524075495
  • The Psychological CorporationWechsler Test of Adult Reading (computer program)San Antonio, TX, USAThe Psychological Corporation2001
  • LovibondPFLovibondSHThe structure of negative emotional states: comparison of the Depression Anxiety Stress Scales (DASS) with the Beck Depression and Anxiety InventoriesBehav Res Ther1995333353437726811
  • BentonALHamsherKMultilingual Aphasia ExaminationIowa City, IA, USAUniversity of Iowa1976
  • SpreenOStraussEA Compendium of Neuropsychological Tests: Administration, Norms, and Commentary2nd edNew York, NY, USAOxford University Press1998
  • RuffRMLightRHParkerSBLevinHSBenton Controlled Oral Word Association Test: reliability and updated normsArch Clin Neuropsychol19961132933814588937
  • MatthewsCGKløveHInstruction Manual for the Adult Neuropsychology Test BatteryMadison, WI, USAUniversity of Wisconsin Medical School1964
  • KellandDZLewisRFEvaluation of the reliability and validity of the repeatable cognitive-perceptual-motor batteryClin Neuropsychol19948295308
  • MeadorKJMooreEENicholsMEThe role of cholinergic systems in visuospatial processing and memoryJ Clin Exp Neuropsychol1993158328428276939
  • ChervinskyABMitrushinaMSatzPComparison of four methods of scoring the Rey-Osterrieth Complex Figure Drawing Test on four age groups of normal elderlyBrain Dysfunction19925267287
  • LezakMDNeuropsychological Assessment3rd edNew York, NY, USAOxford University Press1995
  • IngramFSoukupVMIngramPTThe Medical College of Georgia Complex Figures: reliability and preliminary normative data using an intentional learning paradigm in older adultsNeuropsychiatry Neuropsychol Behav Neurol1997101441469150516
  • ReyAL’examen clinique en psychologie [The clinical examination in psychology]Paris, FrancePresses Universitaires de France1964
  • TaylorEMThe Appraisal of Children with Cerebral DeficitsCambridge, MA, USAHarvard1959
  • LemaySBédardMRouleauITremblayPPractice effect and test-retest reliability of attentional and executive tests in middle-aged to elderly subjectsClin Neuropsychol20041828430215587675
  • BenjaminiYHochbergYControlling the false discovery rate: a practical and powerful approach to multiple testingJ R Stat Soc Series B Stat Methodol199557289300
  • HowellDCStatistical Methods for PsychologyBelmont, CA, USAWadsworth2013
  • AnsteyKJLowLFNormal cognitive changes in agingAust Fam Physician20043378378715532151
  • CortinaJMWhat is a coefficient alpha? An examination of theory and applicationsJ Appl Psychol19937898104
  • GreenMSKayeJABallMJThe Oregon brain aging study: neuropathology accompanying healthy aging in the oldest oldNeurology20005410511310636134
  • VolkowNDGurRCWangGAssociation between decline in brain dopamine activity with age and cognitive and motor impairment in healthy individualsAm J Psychiatry19981553443499501743
  • WilsonRSLeurgansSEBoylePASchneiderJABennettDANeurodegenerative basis of age-related cognitive declineNeurology2010751070107820844243
  • ElwoodRWMicroCog: assessment of cognitive functioningNeuropsychol Rev2001118910011572473
  • RobbinsTWJamesMOwenAMA study of performance on tests from the CANTAB battery sensitive to frontal lobe dysfunction in a large sample of normal volunteers: implications for theories of executive functioning and cognitive agingJ Int Neuropsychol Soc199844744909745237
  • ForsterKIDavisCRepetition priming and frequency attenuation in lexical accessJ Exp Psychol Learn Mem Cogn198410680698
  • MattinglyJBRichANYellandGBradshawJLUnconscious priming eliminates automatic binding of colour and alpha-numeric form in synaesthesiaNature200141058058211279495
  • SpeirsSYellandGWRinehartNTongeBLexical processing in individuals with high-functioning autism and Asperger’s disorderAutism20111530732521363869
  • BleeckerMLFordDPLindgrenKNHoeseVMWalshKSVaughanCGDifferential effects of lead exposure on components of verbal memoryOccup Environ Med20056218118715723883
  • PapazoglouAKingTZMorrisRDMorrisMKKrawieckiNSAttention mediates radiation’s impact on daily living skills in children treated for brain tumorsPediatr Blood Cancer2008501253125718260121
  • MaruffPThomasECysiqueLValidity of the CogState brief battery: relationship to standardized tests and sensitivity to cognitive impairment in mild traumatic brain injury, schizophrenia, and AIDS dementia complexArch Clin Neuropsychol20092416517819395350
  • GreenRCGreenJHarrisonJMKutnerMHScreening for cognitive impairment in older individualsArch Neurol1994517797868042926