11,365
Views
54
CrossRef citations to date
0
Altmetric
Articles

Development of the International Ocean Literacy Survey: measuring knowledge across the world.

ORCID Icon, , &
Pages 238-263 | Received 13 Oct 2017, Accepted 09 Feb 2018, Published online: 12 Mar 2018

Abstract

The Ocean Literacy movement began in the U.S. in the early 2000s, and has recently become an international effort. The focus on marine environmental issues and marine education is increasing, and yet it has been difficult to show progress of the ocean literacy movement, in part, because no widely adopted measurement tool exists. The International Ocean Literacy Survey (IOLS) aims to serve as a community-based measurement tool that allows the comparison of levels of ocean knowledge across time and location. The IOLS has already been subjected to two rounds of field testing. The results from the second testing, presented in this paper, provide evidence that the IOLS is psychometrically valid and reliable, and has a single factor structure across 17 languages and 24 countries. The analyses have also guided the construction of a third improved version that will be further tested in 2018.

1. Introduction

The ocean covers 71% of our planet and holds 97% of the Earth’s water. It is a key ecosystem that encompasses most of the living space on Earth and plays several crucial roles that support the health of the planet and the livelihood of humans. The ocean provides about 15% of the total protein consumed by people across the globe (World Health Organization Citation2012), drives a substantial portion of the global economy (OECD Citation2016), regulates the climate and weather, and slows climate change by absorbing about 40% of the carbon dioxide that is being emitted into the atmosphere at an increasing pace by human activities since the beginning of the industrial revolution (DeVries, Holzer, and Primeau Citation2017). Clearly, the ocean supports life on Earth and provides us with tremendous economic, social, and environmental benefits. Moreover, the ocean is not solely a resource for humans, but has intrinsic value for its own sake and for its inhabitants.

Despite its value, the ocean is showing significant signs of change as a result of human activities. Average sea surface temperatures are rising; the chemistry of the ocean itself is changing, which impacts marine ecosystems and their services (Pörtner et al. Citation2014); many commercially important fish stocks are fully exploited, overexploited, depleted or recovering from depletion, putting marine biodiversity at risk; and the increasing environmental, social and economic pressures from the exploding human population have led to massive alteration of marine habitats (Rockström et al. Citation2009). The increasing modification, degradation and contamination of the ocean directly threatens humankind by putting at risk many associated goods, services, and aesthetic and spiritual benefits. Since this has direct impact on communities and nations worldwide, and can be attributed to the lifestyles, decision-making, and choices of individuals, as well as, governments and industry, the involvement of each and every person in understanding the importance of the ocean and the need to protect it are essential (Fauville Citation2017). For individuals to become thoughtful participants in the debate about solutions to marine environmental issues, they need to be ocean literate.

While the primary meaning of the concept of literacy solely refers to the ability to read and write, this concept has evolved through time. Its meaning has been extended to include the ability to understand a text and be able to make sense of and use it in the world for relevant purposes (Wertsch Citation1991). More recently, UNESCO expanded the concept of literacy by stating that, “Literacy involves a continuum of learning in enabling individuals to achieve his or her goals, develop his or her knowledge and potential, and participate fully in community and wider society.” (UNESCO Citation2005, 21)

Various types of literacy, such as science literacy, digital literacy, environmental literacy or ocean literacy point to skills that are essential in our time and that include but go beyond reading and writing in the classical sense.

Cava et al. (Citation2005) defined Ocean Literacy as an understanding of the ocean’s influence on us and our influence on the ocean. Elaborating on this understanding of interdependencies, the authors define an ocean literate person as someone who understands the essential principles and fundamental concepts about the functioning of the ocean, is able to communicate about the ocean in meaningful ways, and is able to make informed and responsible decisions regarding the ocean and its resources.

Ocean Literacy is aligned with the objectives of environmental education as defined by UNESCO in 1975:

Awareness: to help social groups and individuals acquire an awareness of and sensitivity to the global environment and its allied problems.

Attitude: to help social groups and individuals acquire a set of values and feelings of concern for the environment, as well as, the motivation to actively participate in environmental improvement and protection.

Skills: to help social groups and individuals acquire the skills for identifying and solving environmental problems.

Participation: to provide social groups and individuals with an opportunity to be actively involved at all levels in working towards resolution of environmental problems. (UNESCO (United Nations of Education Scientific and Cultural Organisation) Citation1975, 26–27).

Moreover, according to the National Research Council (Citation2010), which reviewed the education programs of the U.S. National Oceanic and Atmospheric Administration, ocean sciences education, as a means to promote Ocean Literacy, is situated at the intersection of environmental education and STEM (Science, Technology, Engineering and Mathematics) education.

Previous research from several countries has shown that citizens have a limited understanding of marine-related phenomena (Brody Citation1996; Fortner and Mayer Citation1991; Guest, Lotze, and Wallace Citation2015), hold misconceptions (Ballantyne Citation2004), and/or have little understanding of marine environmental issues and protection (Eddy Citation2014). This lack of familiarity with the ocean can be associated with the fact that ocean concepts are rarely represented in the formal science education curriculum ( Fauville et al. Citationforthcoming; Hoffman, Martos, and Barstow Citation2007). This omission of ocean related topics triggered grass roots and policy-driven responses aimed at giving the ocean its legitimate central role in science and environmental education.

The grass roots movement for Ocean Literacy started in 2002 in the United States with concerned scientists, and formal and informal educators raising their voices to include ocean sciences in the school curriculum. This resulted in a two-week online workshop and extensive follow-up dialogue between hundreds of U.S. ocean sciences and education stakeholders in 2004 (Cava et al. Citation2005). This community discussed what knowledge citizens should master by the end of Grade 12 in the U.S. to be considered ocean literate (Schoedinger, Tran, and Whitley Citation2010), and to build consensus that Ocean Literacy is an essential component of science literacy (Strang, Schoedinger, and de Charon Citation2007). This process resulted in seven overarching ideas called the essential principles of Ocean Literacy and 44 fundamental concepts (In the 2013 revision, an additional concept was added, resulting in the current total of 45) that elaborate each principle (Figure ). This ‘ocean literacy framework’ was originally published as, Ocean Literacy: The Essential Principles and Fundamental Concepts of Ocean Sciences Grades K-12 (National Geographic Society et al. Citation2005), revised in 2013 (National Oceanic and Atmospheric Administration Citation2013), and supplemented by The Ocean Literacy Scope and Sequence for Grades K-12 (National Marine Educators Association Citation2010). Rather than serving as a comprehensive ocean sciences curriculum, the Ocean Literacy principles and concepts serve to answer the question, ‘what ideas about the ocean are so fundamental and important that if students did not understand them, they could not be considered science literate?’

Figure 1. The seven Essential Principles of Ocean Sciences.

Figure 1. The seven Essential Principles of Ocean Sciences.

Since 2004, there has been a growing effort to improve Ocean Literacy around the world (Dupont and Fauville Citation2017). The U.S. National Science Foundation invested over $40 M over a 12 year period in a network of Centers for Ocean Sciences Education Excellence, and the European Union invested more than 7 M Euro in two large, multi-year international projects, SeaChange and ResponSEAble. The U.S. National Oceanic and Atmospheric Administration is currently collaborating with Canada and the European Union on a Transatlantic Ocean Literacy initiative. New professional organizations and networks, similar to the longstanding U.S. National Marine Educators Association, have emerged, including the International Pacific Marine Educators Network, the European Marine Science Educators Association, the Canadian Network for Ocean Education and the Asia Marine Educators Association. All of these efforts have the stated objective of improving Ocean Literacy. Despite these increased investments in Ocean Literacy, it has been difficult to show progress of the Ocean Literacy movement, in part, because no widely adopted measurement tool exists. Previous researchers on ocean knowledge have used a wide variety of methods, target groups, content goals, and conceptual frameworks. The need for a shared measurement tool has been expressed by members of the Ocean Literacy community around the world to determine the impact of particular interventions, to establish a baseline of Ocean Literacy in particular communities, to detect in change in Ocean Literacy levels in particular communities over time, and to compare differences in levels of Ocean Literacy across communities.

The International Ocean Literacy Survey (IOLS), presented in this paper, aims to serve as a community-based measurement tool that allows the comparison of levels of ocean knowledge across time and location. Community-based in this context refers to two things: (1) The IOLS is being developed on a voluntary, grass roots basis by members of the Ocean Literacy community, and (2) While initial testing of the IOLS is being conducted on a national level for the purpose of validating the instrument in a variety of languages and populations, we anticipate that the finished survey will be most widely used at the community level.

In response to the considerable international need, the lack of funding sources for international collaborations, and the enthusiasm of the Ocean Literacy community, the authors have marshaled the contributions of dozens of organizations, institutions, investigators, and practitioners to engage in a somewhat non-traditional, iterative, community-based research design. We have assembled a survey instrument that has been subjected to two rounds of field testing (the first in English in the U.S., the second in 17 languages in 24 countries), has yielded promising results, and is poised for a third international test. We envision that this paper is the first installment in a series that will tell the story of our efforts to create a nimble yet rigorous process for conducting research that is immediately helpful to both practitioners and investigators. In addition to measuring levels of Ocean Literacy around the world, we also intend to inform other large scale international research-based collaborations.

2. Methodology

Ocean Literacy includes three dimensions: knowledge, communication, and decision-making. These three dimensions represent approximations, stated in lay terms for public and practitioner audiences, of the accepted objectives of environmental education described by UNESCO (Citation1975) and of environmental literacy described by the North American Association of Environmental Education (NAAEE Citation2011). The IOLS currently focuses on measuring knowledge as a first step toward a more integrated set of measurement tools addressing each aspect of ocean literacy. We are fully aware that levels of knowledge about the ocean do not alone correlate or lead to all three dimensions of ocean literacy. Two signature challenges associated with this project are that (1) its success depends on collaboarion and cooperation among dozens of disparate members of the marine education community from many countries, cultures and linguistic groups, and (2) that the project is as yet largely un-funded. Given these challenges, we made a strategic decision to begin our efforts by focusing on knowedge as the area where there is broad agreement about the content framework (National Oceanic and Atmospheric Administration Citation2013) that constitutes the foundation of our work across the field.

The IOLS has been comprised of a series of multiple-choice questions addressing all seven principles and most of the 45 concepts of Ocean Literacy (future versions will address all 45). Since these principles and concepts were defined by the Ocean Literacy community as what students should know by the end of high school, the target audience for the IOLS is 16–18 year old students. This specific age range was selected so that we could capture a comparable sample of youth near the end of their compulsory education across variations in science course taking both within and across countries. The Ocean Literacy Scope and Sequence for Grades K-12 (National Marine Educators Association Citation2010), especially the section related to Grades 9–12 (equivalent to ages 14–18), provides a much more detailed and developmental set of sub-concepts that lead to full understanding of the seven principles and 45 concepts. Assessing understanding of the entire Ocean Literacy Scope and Sequence in a survey such as the IOLS would require significantly more items and would be impractical. For our purposes, it is most important to assess students’ understanding of the terminal principles and concepts, and less important to assess students’ developmental progress toward understanding them. Therefore, the items in the IOLS refer to the Ocean Literacy principles and concepts. There are not an equivalent number of items for each of the seven principles since each principle represents different amounts and depths of knowledge.

As a first step in the community-wide participation in the development of an open-source instrument, researchers contributed previously developed whole surveys or individual multiple-choice items to the IOLS project (COSEE unpublished work; Dromgool, Burke, and Allard Citation2015; Greely Citation2008; Guest Citation2013; Robinson and Sardo Citation2015). These items were compiled, reviewed, culled for redundancy and/or edited. In addition, many new items were generated by a team led by the authors, as well as, members of the National Marine Educators Association, and several volunteer ocean scientists from several countries. A pilot study was conducted in June 2016 with an initial set of 50 survey items that were administered to 417 U.S. 16 to 18 year old students using the online survey software Qualtrics. Respondents were recruited from existing networks of teachers with a special interest in education about the ocean. This pilot study helped us to identify problematic items that were, for example, outside of the range of appropriate difficulty (too easy or too hard) or appeared to have responses that were driven by something other than Ocean Literacy (e.g., reading comprehension). Based on these results, we revised the items to create a more cohesive instrument that better aligned with the concepts of Ocean Literacy.

The second version of the IOLS is the focus of this study. It included 48 questions aligning with most of the Ocean Literacy principles and concepts (see Appendix 1). For example as presented in Appendix 1, Q6 (How is sea level measured? A. Average depth of the ocean. B. Average height of the ocean relative to the land. C. Level of the ocean at the lowest tide. D. Level of the ocean at the highest tide.) aligns with Concept 1d

Sea level is the average height of the ocean relative to the land, taking into account the differences caused by tides. Sea level changes as plate tectonics cause the volume of ocean basins and the height of the land to change. It changes as ice caps on land melt or grow. It also changes as sea water expands and contracts when ocean water warms and cools.

The IOLS was originally designed in English. Due to enthusiasm in the Ocean Literacy community it was translated by volunteer researchers into 16 languages (Catalan, Croatian, Danish, Dutch, French, German, Greek, Italian, Japanese, Norwegian, Polish, Portuguese, Simplified Chinese, Spanish, Swedish, and Traditional Chinese). This process of translation served two purposes: to create versions of the instrument for use in other languages and to function as a systematic review of the items themselves. In the absence of being able to conduct cognitive interviews with students from each of the countries and representing each of the languages tested, we invited the translators to provide feedback on the items themselves; specifically, the ocean science content, clarity of the wording, and potential complexities introduced by the translation process.

At the end of August 2016, The IOLS V2 was made available on the online software Qualtrics. The authors made use of a wide range of mailing lists, private and professional social media platforms, and newsletters to invite educators to share the survey with their colleagues and to administer it to their 16–18 year old students around the world in the appropriate language for the population. Between August 2016 and October 2016, 6871 individuals agreed to be in the study.

3. Data analyses and results

Since not all questions on the assessment are ‘questions’ (e.g. fill in the blank) we use the term ‘item’ to refer to a combination of a ‘prompt’ and ‘response options.’ Response options refers to the choices from which respondents could select their response to the prompt. The response data are the response options chosen by respondents. Response data for all items were transformed into dichotomously scored data for psychometric analyses, that is, the responses were scored as incorrect (0) or correct (1) for each of the items. In some cases, to correctly respond to the prompt, the respondents needed to select more than one response option for a particular item. In these cases, each response option is scored separately, either responded to correctly or incorrectly and is treated as a separate item. This led to 79 unique items in the IOLS data.

Data were analyzed using the Rasch model (Rasch Citation[1960] 1980) within the Item Response Theory (IRT) modeling framework. Mathematically, the probability of a correct response in the Rasch model can be expressed as

P(x/θj)=exp(θjβi)1+exp(θjβi)

where a response vector is represented by x = (x1,…, xi). Beta i is the difficulty parameter for item i, and theta j is the ability parameter for a respondent j. P represents the probability of a correct response to item i by a respondent j with ability theta. In this case, this model assumes that a respondent’s ability, ocean literacy knowledge, and the difficulty of the item (i.e. a fixed difficulty relative to the other items, not relative to the ability of the respondent) are the only factors that will influence whether or not the individual gets the item correct.

We performed a series of psychometric analyses to examine the measurement quality of the Ocean Literacy scale. We checked the assumption of the Rasch model that the items measure an underlying unidimensional trait, ocean literacy. Further, we examined the item characteristics, including item difficulty, item characteristic curve [ICC]), reliability, and the quality of the distractors (i.e. incorrect answer options). We also performed fit assessments to detect whether the set of items are consistent with the Rasch model at both the model and item levels; as well as differential item functioning (DIF) to ensure the items are functioning equivalently across subgroups (e.g. gender) in the data. The DIF analyses were implemented in DIF Analysis Software (DIFAS; Penfield Citation2005) and the rest of the psychometric analyses were implemented in R (Mair and Hatzinger Citation2007; R Core Team Citation2017; Revelle Citation2017; Yves Citation2012).

3.1. Descriptive statistics

Table displays descriptive statistics of our study sample, showing gender and language for the survey. As can be seen in the table, many of the survey responses were either from Taiwan and completed in the survey in Traditional Chinese or from the United States and completed the survey in English. No other single country or language had a comparable sample size to these two, however, when taken as a whole, the survey responses across the countries and languages of Europe provide sufficient sample size for comparison to the U.S. English and Taiwanese samples. Within each country and language category we had comparable participation of males and females. In no way do we argue that our study sample is representative of the world, nor is the sample representative of the country and linguistic groups they are drawn from. Although effort was given to recruit samples in relatively similar ways across participating locations, the recruitment process was not uniformly systematic across countries, nor were they randomly drawn from a population. That is, these samples of convenience within each country do not reflect the overall population in that country. Therefore observed differences in scores are just as likely, if not more likely, due to variations in recruitment of the sample than variation in levels of Ocean Literacy of the population in those countries. This is an important limitation to possible inferences from these data.

Table 1. Descriptive statistics of the study sample.

3.2. Dimensionality

IRT models, including the Rasch model, assume the items forming the scale are unidimensional, which means only a single latent trait (Hambleton, Swaminathan, and Rogers Citation1991), Ocean Literacy, drives the responses to the set of items. In this study, we tested this unidimensionality assumption using a Confirmatory Factor Analysis (CFA) with weighted least squares estimator, which is a robust estimator and allows for modeling categorical or ordinal data.

Unidimensionality assumption was evaluated via CFA. Rules of thumb (see Brown Citation2015; Hu and Bentler Citation1999) were a cutoff value close to 0.95 for Tucker-Lewis Index (TLI) and Comparative Fit Index (CFI), a cutoff value less than 0.08 for Standardized Root Mean Square Residual (SRMR), and a cutoff value less than 0.06 for Root Mean Square of Approximation (RMSEA); or using a combination of SRMR less than 0.09 and RMSEA less than 0.06; would generate lower Type II error rates with acceptable Type I error rates. Results of all CFA model fit indicators met the criteria (see Hu and Bentler Citation1999) indicating the IOLS scale did not violate the assumption of unidimensionality. The results indicate that the set of knowledge data fit the one-factor model well (i.e. χ2 = 21923.61, df = 3002, p < .05; Comparative Fit Index [CFI]=0.919; Tucker-Lewis Index [TLI]=0.917; Root Mean Square of Approximation [RMSEA]=0.036, Standardized Root Mean Square Residual [SRMR]=0.038). All knowledge items, except for two (i.e. Q40, Q26_4, see Appendix 1) generated significant factor loadings to the single factor. This implies that all but these two items were psychometrically relevant to measuring Ocean Literacy. We have since revised both of these items for version 3 of this assessment (see Appendix 1 for revisions).

3.3. Reliability

The traditional way to determine reliability is to use Cronbach’s alpha. However, Cronbach’s alpha assumes that the underlying data are continuous variables, and in our case the data are coded as dichotomous (correct or incorrect). To account for the non-continuous underlying data we used the polychoric matrix for the internal consistency estimation and computed the ordinal reliability (Zumbo, Gadermann, and Zeisser Citation2007). The ordinal reliability of the knowledge scale was excellent (alpha = 0.94), indicating that the knowledge scale was a well-defined construct – Ocean Literacy. Equally, all items positively contributed to the overall scale reliability.

3.4. Model fit and item fit

Model fit and item fit statistics include INFIT and OUTFIT indices. OUTFIT detects unexpected responses to items with a difficulty distant from a person’s ability level (Linacre Citation2002); for example, OUTFIT is high when several low ability respondents correctly answer a difficult item or when high ability respondents get a relatively easy item incorrect. INFIT, on the other hand, detects responses to items that are so aligned to a person’s ability level that the item provides redundant information to the other items on the scale. In this study, we calculated Mean-square statistics (MNSQ) and, aligning with convention, considered items as potentially misfitting if their MNSQ values were smaller than 0.5 or larger than 1.5 (Linacre Citation2002).

The results of model and item fit statistics are listed in Table . Overall, the average values of INFIT and OUTFIT statistics were nearly perfect: 0.99 and 0.97, respectively. This means that our data fit the Rasch Model very well. Among the set of items, one item (i.e. Q47_1) had OUTFIT MNSQ value as 1.51, just above our threshold of fit value. This item was flagged as potentially misfitting and requiring further review.

Table 2. Psychometric properties of each scored response in the same order they were presented to survey respondents.

3.5. Item characteristics

In the Rasch model, item and ability parameters are aligned on the same latent trait continuum or scale. The set of IOLS knowledge items had Rasch item difficulty ranging from -2.77 to 2.35, with a mean difficulty value of zero (see Table ). Among these 79 items, 35 of them had Rasch difficulty measures above zero while 44 of them had difficulty measures below zero. Figure is an example of ICC plots. In ICC the probability as a function of ability forms a logistic S-shaped curve, in which the vertical axis is the probability of getting an item correctly, and the horizontal axis is the scaled units of the latent trait. A respondent with higher ability levels on the latent trait (i.e. Ocean Literacy) would have higher probability of getting a correct response, hence the vertical axis increases as the horizontal axis increases. Using Rasch, ICCs across all items have the same slope but vary by their locations (i.e. difficulties) on the latent trait (i.e. Ocean Literacy). In ICC, ‘location describes the extent to which items differ in probabilities across trait levels’ (Embreston and Reise Citation2000). Figure is a person-item map, which compares the distribution of ability for the respondents with the item difficulty for the scales. The person-ability distribution is shifted to the right of the center of the item-difficulty distribution. This implies that abilities were higher than the item difficulties. Said another way, the items were easy for respondents in our study sample. The least difficult item in the instrument was Q2 (also seen in the ICC plot below). This question asked:

Which statement is true: Q2_1 The ocean covers 70% of the Earth’s surface; Q2_2 The land covers 70% of the Earth’s surface; Q2_3 The ocean and the land each cover 50% of the Earth’s surface; Q2_4 The ocean covers 10% of the Earth’s surface

Over 90% of respondents answered this item correctly (Q2_1), indicating a very easy item. The revision of this item is described in the discussion section, and listed in Appendix 1. The most difficult item in the instrument was Q30. This question asked: The accumulation of oxygen in Earth’s atmosphere was necessary for life to develop and be sustained on land. Where did this oxygen originate? Q30_1 Oxygen was already there when the Earth was formed. Q30_2 All oxygen originated from photosynthetic organisms on land. Q30_3 All oxygen originated from photosynthetic organisms both on land and in the ocean. Q30_4 All oxygen originated from photosynthetic organisms in the ocean. This item has been modified to reduce the reading demand and improve the overall clarity of the item (See Appendix 1 for the changes to the item, and see Appendix 2 for Item difficulty estimates).

Figure 2. Examples of some ICC plots.

Figure 2. Examples of some ICC plots.

Figure 3. Person-Item Map.

Figure 3. Person-Item Map.

3.6. Distractor analysis

We conducted the distractor analysis to determine whether item options function effectively. Desirable distractors should attract respondents to choose them when respondents are unsure of the correct answer but should not be so attractive that respondents choose them more often than the correct answer.

In this study, we examined the frequency distribution of item options chosen by respondents; any item option that was chosen less than 5% of the time was flagged for follow up discussion and potential revision. Results of distractor analyses indicate that some of the items (identified in Appendix 1 with the symbol ‘Ψ’) had options that were not sufficiently attractive to respondents. For example, for item Q2, only 1.9% of respondents chose the second option, 1% chose the third option, and 0.5% chose the fourth option; meaning that these distractors were not providing useful information to differentiate low and high proficient respondents; 96.6% of respondents selected the correct answer (i.e. the first option) to this item. These item options would benefit from a thoughtful revision. Appendix 1 provides the complete list of item options that were suggested for content expert review.

3.7. Differential item functioning

Differential item functioning (DIF) means that items function differently across sub-groups in the sample. One example of DIF is when a boy and a girl that have the same ability estimates, but have different probabilities of getting the item correct because the item is easier for one gender than it is for another. DIF analysis is essential in the development of a scale in order to determine if the test is fair across respondents.

In this study, DIF detection was implemented with nonparametric analyses for dichotomously scored items. This study used the Mantel-Haenszel chi-square (Holland and Thayer Citation1988; Mantel and Haenszel Citation1959) and Educational Testing Services (ETS) classification scheme for evaluating DIF (i.e. A = negligible DIF; B = moderate DIF; C = large DIF; Zieky Citation1993). The criteria to diagnose a DIF item in this study is the presence of both statistical significance (i.e. Mantel-Haenszel chi-square value exceeding 3.84, or p < 0.05) and practical significance (i.e. the presence of moderate or large levels of the ETS DIF classification scheme) (Chen and Jiao Citation2014).

In implementation, this study used the summated test score of the items selected for the DIF analyses as the stratification variable. The grouping variables included (1) boy and girl; (2) Taiwan and non-Taiwan; and (3) Europe and USA. These comparisons were chosen because each had sufficient sample size and each provided insight on test characteristics. Examining DIF across gender allows us to use all the data collected from around the world to look for differences in gender responses. Taiwan and non-Taiwan is important to examine since the Taiwan data make up such a large portion of the total data, we wanted to ensure that their responses are not skewing the overall results. Europe and the USA allows for vetting the instrument across these cultural and linguistic differences. Table summarizes the flagged potential DIF items. Results show that three items were flagged as potentially showing DIF across gender (i.e. Q3, Q8, Q4_1). Further, 40 items were flagged as having potential DIF effects between Taiwan and non-Taiwan respondent groups. A common characteristic across many of the items that favor Taiwanese students is that these items contained words such as ‘never’ or ‘always’ in some of the incorrect response options. These items were easier to answer correctly for respondents in Taiwan compared to non-Taiwanese students. One possible explanation for this DIF effect is that Tawainese students learn specific test-taking strategies; for example, eliminating response options with these words. Revising these items to eliminate these words is in alignment with best practices for assessment construction and may eliminate much of the DIF observed between Taiwan and non-Taiwan students. Also, further screening from content experts would be needed to see whether there is any translation issue that could contribute to these differences. Results show that 22 of the items potentially had DIF effects between Europe and the USA. While the sheer number of items that demonstrated DIF was high, that does not mean that all of these items were differentially functioning. The analysis merely points to items that function differently across different sub-groups, but doesn’t tell us why. While much of this difference maybe due to construct irrelevant differences (e.g., better test takers in Taiwan than in the USA), some of this difference is construct relevant, meaning that the difference is related to the subject of study. For example, item Q18 which was flagged for DIF asked about changing sea levels; respondents in Taiwan, an island nation along the tropic of cancer, had more difficulty selecting ice caps melting and growing as having an influence on sea levels than respondents outside of Taiwan (mostly in the U.S. and Europe). This difference in difficulty may be due to different emphases in the curriculum across these locales given their relative proximity to polar ice caps. This may be the type of difference across countries that the survey is aiming to uncover.

Table 3. Summary of flagged potential DIF items in the same order they were presented to survey respondents.

3.8. Modifications and preparations for version 3

Developing a single instrument that functions equally across linguistic and cultural differences is exceedingly difficult. The results from the second trial are informing the construction of the next version of the survey. Content experts are currently revisiting the items to review them for clarity, content alignment, and explore ways to modify the items to perform better across participating countries and languages. The results of this process can be seen in Appendix 1. Additional items have been added to the assessment to measure concepts that had previously been left out of the survey.

3.9. Summary of the findings

Our analyses indicate that the survey indeed assesses one factor, that we are referring to as ‘Ocean Literacy.’ Given the challenges associated with the community-development of a survey in 17 languages, this provides encouragement to continue development of this international collaborative effort. Because the survey was initially developed by gathering items from existing instruments, and only a few previous studies existed, we were limited in the types of items included. Many of the items, for instance, assessed only declarative knowledge or factual recall. Others had inconsistent construction of distractors, with spurious words, inconsistent distractor length, or contain words like ‘never’ or ‘always’ that often indicate that these are not the correct answer. Appendix 1 shows how we have modified many questions from V2, discussed in this paper, to create V3, to minimize these issues and which will be administered for a new round of testing.

Some modifications are intended to simply make small improvements to an item based directly on analysis of results from V2, i.e., the item did not test well either because one or more of the distractors were infrequently selected, or the item favored a particular population for reasons we think are unrelated to understanding of Ocean Literacy. For example, in Q1 ‘Which statement is the most accurate,’ some distractors are negative statements while some are positive, and distractors Q1_1 (The water in the Pacific Ocean will never reach the Indian Ocean) and Q1_4 (The water in the Gulf of Mexico can never reach the Pacific Ocean) both contain the word, ‘never.’ The item favored Taiwanese respondents, who may learn to avoid distractors that include words such as always and never. We have revised the item for V3 so that the correct response and all distractors are positive statements, and we eliminated the use of the words, ‘never.’

Other modifications are intended to reframe items from representing declarative knowledge to more conceptual understanding. For example, Q2 ‘Which statement is true: Q2_1 The ocean covers 70% of the Earth’s surface; Q2_2 The land covers 70% of the Earth’s surface; Q2_3 The ocean and the land each cover 50% of the Earth’s surface; Q2_4 The ocean covers 10% of the Earth’s surface’ asks respondents only to recall an important fact about the ocean. In V2, 96.6% of respondents answered this question correctly. The question did not provide useful information to differentiate low and high respondents. The concept that most of Earth’s surface is covered by the ocean is a defining idea in Ocean Literacy, but respondents’ ability to recall of the percentage does not indicate understanding of why this idea is so important to earth systems. Q2 has been revised for V3 to be more conceptual in nature:

Because the ocean covers most of the Earth (select the best answer): Q2_1 It controls our weather, climate and oxygen production; Q2_2 Most living things are concentrated on the continents; Q2_3 Lots of the Earth is not very useful for humans; Q2_4 It generates most of the Earth’s greenhouse gases.

3.10. Limitations

The samples used in this study are not representative of the countries from which they are drawn. So, we are careful to not draw conclusions about the level of Ocean Literacy across these countries. However, it is possible that there was systematic bias in the way the data were collected across all countries that lead to a poor representation of students around the world. The systematic bias that would be most harmful for our work would be overly represented knowledge about the ocean. It is possible that our estimates of item difficulty are biased downward because teachers who teach about the ocean were more likely to administer the survey. However, for our purposes the items are being evaluated in a relative sense to each other, not against the sample itself. Therefore, the analyses conducted are relatively robust to any systematic bias in the sample.

4. Discussion

The ocean is an important part of our world, something we all share, we all benefit from and we all have an impact on. Understanding our connection to the ocean and being an informed participant in the discussion of the future of the ocean requires a degree of Ocean Literacy. There have been many considerable investments made over the last 15 years by governments and non-governmental organizations for the purpose of increasing ocean literacy. There have been few attempts, however, to understand the influence of these efforts on learners or the general public. Since the work to improve Ocean Literacy is relatively new, especially outside of the United States, it has focused mainly on programmatic activities and interventions. These interventions until recently, have not attracted much attention from the education research community. The International Ocean Literacy Survey is among the first tools intended to support the efforts of those aiming to educate our citizenry about the ocean that has been translated widely, pilot tested multiple times and subjected to rigorous psychometric scrutiny.

The work presented here provides evidence that the survey instrument is psychometrically valid and reliable, and has a single factor structure across 17 languages and 24 countries. Further, we argue that there is still much work to be done to produce an instrument that can be used equivalently across these different cultural and linguistic contexts. The authors are continuing these efforts. We have made several changes to individual items as a result of our findings, and we will be further testing the Survey in early 2018.

Further, this paper attempts to demonstrate that a small group of people can: lead an international effort on a limited budget with contributions from dozens of researchers and practitioners around the world; maintain the integrity of the research design despite using somewhat non-traditional methods; and make a useful and practical impact on efforts to understand and improve education efforts around the world. We are experimenting with new methods of grassroots, stone soup-style, community-based instrument development, using a process similar to the collective impact framework (Kania and Kramer Citation2011), that relies on a network of committed individuals and organizations, with a common goal and common measures, and led by a trusted backbone organization. The community is willing to forego some traditional academic needs for ownership and authorship in order to achieve practical collective results that mark steady progress toward achieving the goal. Dozens of researchers, evaluators, scientists and educators freely contributed original instruments or individual items, edited or adapted items, reviewed or administered the survey, and advised on the process. The large number of contributors posed some challenges to the research design, but in the end, all contributors have a single goal, to assist in the development of a new, universal instrument that allows those in the community to measure progress and compare results across user groups. The goal is to create an instrument that represents a whole that is greater than the sum of its parts. We invite additional investigators interested in either the methods and technical aspects of the effort or in the advancement of understanding about Ocean Literacy to participate in the ongoing development and administration of the International Ocean Literacy Survey.

The International Ocean Literacy Survey is designed to detect progress toward, and so, to support the improvement of, international and potentially global efforts to build public understanding of the importance of the ocean. Use of the IOLS is a key strategy for justifying and promoting efforts to increase the public’s capacity to understand, communicate about, manage, sustain and protect ocean resources and ocean ecosystems.

We recognize the essential need to move beyond measuring only ocean knowledge to really understand levels of Ocean Literacy among our 16–18 year old audience. Ocean Literacy is defined as also including the ability to communicate about the ocean in a meaningful way, and to be able to make informed and responsible decisions regarding the ocean and its resources. Knowledge of the essential principles and fundamental concepts about the ocean is the dimension of Ocean Literacy most well defined, so we have chosen to begin our measurement efforts there. We intend to expand our efforts over time to include measures of communication and decision-making.

Disclosure statement

No potential conflict of interest was reported by the authors.

Notes on contributors

Géraldine Fauville is a researcher at the Department of Education, Communication and Learning at the University of Gothenburg, Sweden. Her studies focus on the role and potential of digital technologies for contributing to the process of learning about marine environmental issues. She also has a background in marine biology. She has also worked for 10 years as manager of a long-lasting project between Stanford University and University of Gothenburg developing digital learning resources for marine environmental education. She co-founded the European Marine Science Educators Association (EMSEA) in 2011 where she currently serves as Director. She also serves on the Board of Directors of the U.S. National Marine Educators Association (NMEA).

Craig Strang is an associate director, Lawrence Hall of Science at the University of California, Berkeley. He leads the Learning and Teaching Group, and is principal investigator of BaySci: The Bay Area Partnership for K-12 Science; an NSF-funded research study on science professional development, and is PI of a project to improve instruction in outdoor science programs. He is founding director of MARE: Marine Activities, Resources & Education, a K-8 professional learning and curriculum program. He was PI of the NSF-funded Center for Ocean Sciences Education Excellence--California. He was the president of the National Marine Educators Association in 2012, and co-leads the Ocean Literacy Campaign and development of the Ocean Literacy Framework. He is co-author of three sets of science curriculum materials for grades K-8, and is PI of the Ocean Sciences Curriculum Sequences for Grades 3-5 and 6-8. Before turning to science education he did research on elephant seals, humpback whales and California sea lions.

Matthew A. Cannady is the director of Quantitative Studies for the Research and Impact Group at the Lawrence Hall of Science at the University of California, Berkeley. He has been involved in science research and education for more than 15 years. After spending a year as a product design engineer, and planetary scientist at NASA Ames Research Center, he taught high school physics for five years before earning his doctorate degree in Educational Research, Measurement and Evaluation from Boston College. Since then, he has collaborated on numerous large-scale research and evaluation studies in the field of education, including evaluations of teacher professional development, design-based research for after-school science programs, and construction of research measurement instruments.

Ying-Fang Chen is a research specialist at the Research and Impact Group at the Lawrence Hall of Science at the University of California, Berkeley. She brings her expertise in psychometrics, statistics, and evaluation to STEM research projects within the Research and Impact Group. She has many years of professional experience in item response theory modeling, statistical methodology, large-scale survey data systems, as well as educational psychology research and institutional research. She received her doctorate in Measurement, Statistics and Evaluation from University of Maryland, College Park. Her doctoral and postdoctoral research focused on developing innovative psychometric modeling to improve the measurement of large-scale assessments.

Acknowledgment

We would like to acknowledge the essential contributions of the IOLS partner organizations. The following organizations participated in and/or support the development and administration of the IOLS: Asia Marine Educators Association (AMEA), Canadian Network for Ocean Education (CaNOE), National Marine Educators Association (NMEA), Marine Conservation Society, Partnership for Observation of the Global Oceans (POGO), and Surfline. In addition, we thank and acknowledge the following individuals who volunteered countless hours to translate the IOLS into 16 languages:

Chinese Trad: Liang-Ting Tsai, Jack Chang

Croatian: Melita Mokos

Danish: Lene Friis Møller

Dutch: Evy Copejans, Pieter Demuynck

French: Géraldine Fauville

German: Mirjam Glessmer, Maike Nicolai

Greek: Yolanda Koulouri, Maria Cheimonopoulou, Athanasios Mogias, Theodoros Kevrekidis, Theodora Boubonari

Italian: Martina Milanese

Japanese: Tsuyoshi Sasaki

Norwegian: Theodora Sam

Polish: Grażyna Niedoszytko, Weronika Podlesinska, Dominika Wojcieszek

Portuguese: Raquel Costa, Diogo Geraldes, Fernanda Silva

Simplified Chinese: Mo Chen, Guoguang Yang

Spanish and Catalan: Manel Gazo, Carol Campillo Campbell

Swedish: Susan Gotensparre, Björn Källström

Finally, we thank and acknowledge all of the dozens of individuals around the world who contributed previously developed surveys, participated in the development of new items, reviewed and edited the survey for accuracy, and who assisted in the administration of the IOLS to over 6000 students in 24 countries.

References

  • Ballantyne, R. 2004. “Young Students’ Conceptions of the Marine Environment and Their Role in the Development of Aquaria Exhibits.” GeoJournal 60 (2): 159–163.10.1023/B:GEJO.0000033579.19277.ff
  • Brody, M. J. 1996. “An Assessment of 4th-, 8th-, and 11th-Grade Students’ Environmental Science Knowledge Related to Oregon’s Marine Resources.” The Journal of Environmental Education 27 (3): 21–27.10.1080/00958964.1996.9941463
  • Brown, T. A. 2015. Confirmatory Factor Analysis for Applied Research. New York: The Guilford Press.
  • Cava, F., S. Schoedinger, C. Strang, and P. Tuddenham. 2005. “Science Content and Standards for Ocean Literacy: A Report on Ocean Literacy.” Accessed October 13, 2017. http://www.cosee.net/files/coseeca/OLit04-05FinalReport.pdf
  • Chen, Y.-F., and H. Jiao. 2014. “Exploring the Utility of Background and Cognitive Variables in Explaining Latent Differential Item Functioning: An Example of the PISA 2009 Reading Assessment.” Educational Assessment 19: 77–96.10.1080/10627197.2014.903650
  • DeVries, T., M. Holzer, and F. Primeau. 2017. “Recent Increase in Oceanic Carbon Uptake Driven by Weaker Upper-Ocean Overturning.” Nature 542: 215–218.10.1038/nature21068
  • Dromgool, Regan C., N. Burke, and B. Allard. 2015. “Ocean Literacy Knowledge Questionnaire for Teachers – Evaluating Your Knowledge about the Ocean.” Explorers Planning Guide For Primary School Teachers, Explorers Education Programme, Marine Institute, Ireland.
  • Eddy, T. D. 2014. “One Hundred-Fold Difference between Perceived and Actual Levels of Marine Protection in New Zealand.” Marine Policy 46: 61–67.10.1016/j.marpol.2014.01.004
  • Embreston, S. E., and S. P. Reise. 2000. Item Response Theory for Psychologists. Hillsdale, NJ: Lawerence Erlbaum Associates.
  • Fauville, G. 2017. “Digital Technologies as Support for Learning about the Marine Environment: Steps toward Ocean Literacy.” PhD diss., University of Gothenburg, Sweden.
  • Fauville, G., P. McHugh, C. Domegan, Å. Mäkitalo, L. Friis Møller, M. Papathanassiou, C. Alvarez Chicote, et al. Forthcoming. “Using Collective Intelligence to Identify Barriers to Teaching 12–19 Year Olds about the Ocean in Europe.” Marine Policy.
  • Fortner, R. W., and V. J. Mayer. 1991. “Repeated Measures of Students’ Marine and Great Lakes Awareness.” The Journal of Environmental Education 23: 30–35.10.1080/00958964.1991.9943067
  • Greely, T. 2008. “Ocean Literacy and Reasoning about Ocean Issues: The Influence of Content, Experience and Morality.” PhD diss., University of South Florida.
  • Guest, H. 2013. “Connecting Value and Knowledge of Marine Environmental Issues: A Social Survey in Coastal Nova Scotia.” Honours Thesis, Dalhousie University.
  • Guest, H., H. K. Lotze, and D. Wallace. 2015. “Youth and the Sea: Ocean Literacy in Nova Scotia, Canada.” Marine Policy 58: 98–107.10.1016/j.marpol.2015.04.007
  • Hambleton, R. K., H. Swaminathan, and H. J. Rogers. 1991. Fundamentals of Item Response Theory. Newbury Park, CA: Sage.
  • Hoffman, M., D. Martos, and D. Barstow. 2007. Revolutionizing Earth System Science Education for the 21st Century, Report and Recommendations from a 50-State Analysis of Earth Science Education Standards. Cambridge, MA: TERC.
  • Holland, P. W., and D. T. Thayer. 1988. “Differential Item Functioning and the Mantel-Haenszel Procedure.” In Test Validity, edited by H. Wainer and H. I. Braun, 129–145. Hillsdale, NJ: Erlbaum.
  • Hu, L., and P. M. Bentler. 1999. “Cutoff Criteria for Fit Indexes in Covariance Structure Analysis: Conventional Criteria versus New Alternatives.” Structural Equation Modeling 6: 1–55.10.1080/10705519909540118
  • Kania, J., and M. Kramer. 2011. “Collective Impact.” Stanford Social Innovation Review Winter 2011 36–41.
  • Linacre, J. M. 2002. “What Do Infit and Outfit, Mean-Square and Standardized Mean.” Rasch Measurement Transactions 16: 878.
  • Mair, P., and R. Hatzinger. 2007. “Extended Rasch Modeling: The ERm Package for the Application of IRT Models in R.” Journal of Statistical Software 20: 1–20.
  • Mantel, N., and W. Haenszel. 1959. “Statistical Aspects of the Analysis of Data from Retrospective Studies of Disease.” Journal of the National Cancer Institute 22: 719–748.
  • National Geographic Society, National Oceanic and Atmospheric Administration, Centers for Ocean Sciences Education Excellence, and National Marine Educators Association and College of Exploration. 2005. Ocean Literacy: The Essential Principles of Ocean Sciences K-12. Silver Spring, MD: NOAA.
  • National Marine Educators Association. 2010. “Ocean Literacy Scope and Sequence for Grades K-12, Published in the National Marine Educators Association Special Report #3.” In The Ocean Literacy Campaign Featuring the Ocean Literacy Scope & Sequence for Grades K-12. NMEA.
  • National Oceanic and Atmospheric Administration. 2013. Ocean Literacy: The Essential Principles of Ocean Sciences for Learners of All Ages: Version 2: March 2013. Accessed October 13, 2017. http://www.coexploration.org/oceanliteracy/documents/OceanLitChart.pdf
  • National Research Council. 2010. NOAA’s Education Program: Review and Critique. Washington, DC: The National Academies Press.
  • NAAEE (North American Association For Environmental Education). 2011. Developing a Framework for Assessing Environmental Literacy: Executive Summary. Washington, DC: NAAEE.
  • Dupont, S., and G. Fauville. 2017. “Ocean Literacy as a Key toward Sustainable Development and Ocean Governance.” In Handbook on the Economics and Management of Sustainable Oceans, edited by P. Nunes, L. E. Svensson and A. Markandya, 519–537. Cheltenham: Edward Elgar Publishers & UNEP.10.4337/9781786430724
  • OECD (Organisation for Economic Co-operation and Development). 2016. The Ocean Economy in 2030. Paris: OECD Publishing.
  • Penfield, R. D. 2005. “DIFAS: Differential Item Functioning Analysis System.” Applied Psychological Measurement 29: 150–151.10.1177/0146621603260686
  • Pörtner, H. O., D. M. Karl, P. W. Boyd, W. W. L. Cheung, S. E. Lluch-Cota, Y. Nojiri, D. N. Schmidt, et al. 2014. “Ocean Systems.” In Climate Change 2014: Impacts, Adaptation, and Vulnerability. Part a: Global and Sectoral Spects. Contribution of Working Group II to the Fifth Assessment Report of the Intergovernmental Panel on Climate Change, edited by C. B. Field, V. R. Barros, D. J. Dokken, K. J. Mach, M. D. Mastrandrea, T. E. Bilir, M. Chatterjee, et al., 411–484. Cambridge: Cambridge University Press.
  • R Core Team. 2017. R: A Language and Environment for Statistical Computing. R Foundation for Statistical Computing, Vienna, Austria. Accessed October 13, 2017. http://www.R-project.org/
  • Rasch, G. [1960] 1980. Probabilistic Models for Some Intelligence and Attainment Tests. Chicago: The University of Chicago Press.
  • Revelle, W. 2017. Psych: Procedures for Personality and Psychological Research. Evanston, IL: Northwestern University.
  • Robinson, A. W. D., and M. Sardo. 2015. “Making a Splash - the Development and Pilot of a Formal Measure of Public Ocean Literacy.” Master thesis, University of the West of England, Bristol, UK.
  • Rockström, J., W. Steffen, K. Noone, Å. Persson, F. S. Chapin, E. F. Lambin, T. M. Lenton, et al. 2009. “A Safe Operating Space for Humanity.” Nature 461: 472–475.10.1038/461472a
  • Schoedinger, S., L. U. Tran, and L. Whitley. 2010. “From the Principles to the Scope and Sequence : A Brief History of the Ocean Literacy Campaign.” Current: The Journal of Marine Education, Special Report #3 3–7.
  • Strang, C., S. Schoedinger, and A. de Charon. 2007. ““Can You Be Science Literate without Being Ocean Literate?” Current: The Journal of Marine.” Education 23 (1): 7–9.
  • UNESCO (United Nations of Education Scientific and Cultural Organisation). 1975. The International Workshop on Environmental Education Final Report, Belgrade, Yugoslavia. Paris: UNESCO/UNEP.
  • UNESCO (United Nations of Education Scientific and Cultural Organisation). 2005. Aspects of Literacy Assessment: Topics and Issues from the UNESCO Expert Meeting. Paris: UNESCO.
  • Wertsch, J. 1991. “Sociocultural Setting and the Zone of Proximal Development: The Problem of Text-Based Realities.” In Culture, Schooling, and Psychological Development, edited by L. Tolchinsky Landsmann, 71–86. Norwood, NJ: Ablex.
  • World Health Organization. 2012. “Availability and Consumption of Fish.” Accessed October 13, 2017. http://www.who.int/nutrition/topics/3_foodconsumption/en/index5.html
  • Yves, R. 2012. “Lavaan: An R Package for Structural Equation Modeling.” Journal of Statistical Software 48 (2): 1–36.
  • Zieky, M. 1993. “Practical Questions in the Use of DIF Statistics in Item Development.” In Differential Item Functioning, edited by P. W. Holland and H. Wainer, 337–364. Hillsdale, NJ: Erlbaum.
  • Zumbo, B. D., A. M. Gadermann, and C. Zeisser. 2007. “Ordinal Versions of Coefficients Alpha and Theta for Likert Rating Scales.” Journal of Modern Applied Statistical Methods 6: 21–29.10.22237/jmasm/1177992180

Appendix 1

 

Table A1. List of the questions from Version 2 and their modification in V3. Note there is no Q16 and 45 as they were deleted after V1 and the numbers were kept unchanged in order to make it easier to track the questions from one version to the next. Column C. corresponds to the specific ‘fundamental concept’ from the Ocean Literacy Framework that is addressed by each specific question. The items in bold are the correct answers. The symbol ‘Ψ’ represent distractors that were not sufficiently attractive to respondents. The items are placed in numerical order.

Appendix 2

 

Table B1. Item difficulty estimates listed in the same order they were presented to survey respondents (0.95 Confidence Interval).