916
Views
26
CrossRef citations to date
0
Altmetric
Original Articles

Scientific geodata infrastructures: challenges, approaches and directions

, , , &
Pages 613-633 | Received 16 Jul 2012, Accepted 24 Feb 2013, Published online: 10 Apr 2013

Abstract

Based on various experiences in developing Geodata Infrastructures (GDIs) for scientific applications, this article proposes the concept of a Scientific GDI that can be used by scientists in environmental and earth sciences to share and disseminate their research results and related analysis methods. Scientific GDI is understood as an approach to tackle the science case in Digital Earth and to further enhance e-science for environmental research. Creating Scientific GDI to support the research community in efficiently exchanging data and methods related to the various scientific disciplines forming the basis of environmental studies poses numerous challenges on today's GDI developments. The paper summarizes requirements and recommendations on the publication of scientific geospatial data and on functionalities to be provided in Scientific GDI. Best practices and open issues for governance and policies of a Scientific GDI are discussed and are concluded by deriving a research agenda for the next decade.

1. Introduction

It is a bit more than a decade after the former US vice president (Gore Citation1999) coined the term Digital Earth to envisage the future use of virtual (digital) globes based on internet and virtual reality technologies. Gore envisioned Digital Earth as an instrument to gain seamless access to various kinds of globally distributed spatio-temporal datasets each covering different parts of the world, having different scales and resolutions and describing the state of the environment and potential environmental threads. Today, a good part of that vision became reality in a number of mostly commercially driven virtual globe applications that we use on a daily basis to virtually explore places (Goodchild Citation2008). Gore still imagined these applications as being available in only certain institutions, which could provide the required powerful base technology. While the enormous technological progress related to geodata acquisition, computational power, internet protocols and geodata processing allows today for usage of such applications not only on our own desktops but even as ‘apps’ on various kinds of mobile devices. Infrastructures to share geospatial data from spatially distributed and diverse organizations form the backbone of the Digital Earth, and different types of these infrastructures can be found today.

In (re-)defining a vision for a Digital Earth 2020, Craglia et al. (Citation2012) consider the science case to frame a prospective and requirements on future geoinformation technologies supporting earth sciences and environmental research. Taking the progress in information technologies and the on-going developments towards Spatial Data Infrastructures (SDIs), Geodata Infrastructures (GDIs) and the Global Earth Observation System of Systems (GEOSSs) as a starting point, Craglia et al. (Citation2012) identify the following key challenges for the science case in a Digital Earth 2020:

  • linking of multi-disciplinary models to support forecasting and assessing global change(s),

  • integration of (near) real-time observations taken from the fast emerging pervasive modern sensor networks, including social networks,

  • consideration of policy scenarios and their potential impacts,

  • communication of scientific findings on global change effects, the related uncertainties and proposed measures to scientists, decision makers and the public, meanwhile providing participatory frameworks supporting stakeholders in sharing their concerns and formulating responses and actions.

Concurring with these aspects and based on own findings in developing GDI for scientific applications, we propose the concept of a Scientific GDI and address related issues. The paper summarizes experiences which the authors gained in numerous projects related to GDI implementations for various scientific domains. Special mention deserves the GLUES project (Global Assessment of Land Use Dynamics, Greenhouse Gas Emissions and Ecosystem Services) which is the coordination project of the international research program ‘Sustainable Land Management’ of the German Ministry of Education and Research. Within GLUES, the authors implement a scientific GDI in order to facilitate the interdisciplinary data exchange between scientists of the research program (http://modul-a.nachhaltiges-landmanagement.de/en/scientific-coordination-glues/). Additionally, as preparation for this article, an investigation into available scientific data infrastructures has been conducted. In particular, the survey addressed the provided functionalities and contents, the supported data formats and interfaces, system architectures and licensing. The results are incorporated and have affected the overall conclusion of the paper. A summary of the investigated scientific data infrastructures and an approach to classify them can be found here: http://geoportal.glues.geo.tu-dresden.de/scientificinfrastructures/

In the following, a quick review on current GDI developments helps identify bricks and best practices for realizing Scientific GDI. The remainder of the paper summarizes obstacles, requirements and recommendations for the implementation of Scientific GDI and discusses governance and policies to foster Scientific GDI.

2. The status on infrastructures for sharing geoinformation

The idea of establishing internet-based infrastructures for sharing digital geoinformation arose at least two decades ago: The Clinton Order from 1994 marked the legal start of the US National SDI (Masser Citation1999). Meanwhile, a number of initiatives started on establishing GDI and developed common specifications and regulations for geoinformation sharing (). Activities not only address the technological aspects as the Open Geospatial Consortium (OGC) but also provide organizational frames and partly even define first bricks on common semantics to establish the intended Information Infrastructures. The EU Directive ‘Infrastructure for Spatial Information in the European Community’ (INSPIRE; (EP Citation2007a)) is one of Europe's main drivers towards creating an administrative GDI and especially to improve sharing of geoinformation.

Figure 1. Overview on initiatives and infrastructures for sharing geoinformation.
Figure 1. Overview on initiatives and infrastructures for sharing geoinformation.

Consequently, a review on INSPIRE related activities provides a starting point on analysing the current state of GDI, their implementation models and their support in sharing geoinformation. A suite of INSPIRE implementing rules and technical guidance documents (see http://inspire.jrc.ec.europa.eu/) provide the reference frame on which standards should be used and on how spatial data should be made technically available to allow for interoperable applications within the EU. Additional to existing standards, the INSPIRE data specifications define harmonised data models for a number of environmental data themes thus laying the foundation for semantic interoperability and the integration of spatial data from distributed, heterogeneous sources. INSPIRE implementations in the EU member states are on-going and full implementations shall be achieved in 2020. When mainly considering the organizational aspects, a report by De Vries et al. (Citation2011) indicates that INSPIRE clearly boosts national GDI implementations in Europe but that EU member states strategies in implementing INSPIRE vary enormously such that a unified GDI implementing model seems hardly to be observed or achieved. Very recently, the INSPIRE directive together with the planned revision of the European Directive on Public Sector Information (EC Citation2011) encouraged a number of European Member States (e.g. the Netherlands, United Kingdom, Finland, Germany) to at least open the federal official geodata holdings and to provide even topographic data as open data. However, as the real technical INSPIRE implementations are still in their infancy, a valid and thorough report on further effects of INSPIRE cannot be given today.

As an infrastructure, any GDI realization has a requirement to cross-cut administrative and disciplinary boundaries. Additionally, a GDI should follow a service based approach, not only in a technological but also in an organizational sense. Therefore, successfully implementing and operating GDI will most probably necessitate horizontal measures or organizational structures acting as a mediator. The main task of such a mediator – in most cases a new but small institutional – entity is to act highly flexible and to provide all tools and measures that allow an individual public authority to master the tasks being pertinent in implementing GDI.

In terms of functionality, the still emerging first generation of Web-based GDI (GDI 1.0) allows searching for distributed geodata and geoinformation services, interactively visualizing geodata in online available maps, and in best case downloading geodata in a well described format. See for instance Percivall (Citation2010) for a recent overview on related standards. Thus, such GDI 1.0 roughly provide functionality being well comparable to services offered in public libraries: searching on a set of well-defined attributes, browsing and getting a book (or journal, thesis, etc.) if available. Several Geoportals serve as applications to discover and explore the available geodata and geoinformation services (Bernard et al. Citation2005).

Other driving and even pushing factors for today's GDI are the commercial Web-based Map Services (e.g. Google Maps, Bing Maps), the geodata crowd sourcing activities (e.g. OpenStreetMap, GeoNames) and the growing number of available (on-line) geosensors providing (near) real-time observation of various geophenomena (Goodchild and Glennon Citation2010). New combinations of these different information sources are for instance prototyped by the recently launched ‘Eye on Earth’ (http://www.eyeonearth.eu/) and by research on fusion of administrative and crowd sourced geo-information (Wiemann and Bernard Citation2010). It still needs to be investigated to what extent volunteered crowd sourcing activities can offer reliable sources for environmental research, not only in (the obvious) terms of data quality but also in what are the ideal topics, scales and time frames to mobilize a sufficiently sized crowd of volunteers to monitor environmental processes or related effects.

The further improvement of distributed geoprocessing (Müller, Bernard, and Brauner Citation2010) and the usage of cloud computing (Schäffer, Baranski, and Foerster Citation2010; Yang et al. Citation2011) show how future GDI can be enhanced to offer more powerful analysis capabilities. All these activities show a number of organizational and technological components that could be clearly beneficial in implementing next generation GDI and Digital Earth to seamlessly integrate distributed environmental observations and datasets, environmental modelling systems, and processing and analysis functions. Today's GDI that primarily focus on geodata provision could develop into a next generation of service infrastructures offering adequate and user-friendly services to derive the relevant piece of geoinformation and thus getting closer to a Digital Earth directly responding users' information requests.

3. Scientific GDI

Traditionally, the evaluation of scientific work mainly refers to published refereed articles and their impacts. Cleary, this is very often not the only outcome of scientific work. Most scientific activities produce data and methods in the form of software tools which might be valuable beyond their original scope (Gray Citation2009). If scientific results, observations, the underlying simulation models and assumptions are sufficiently described and accessible, they can be valuable input for other scientists and other domains of users. Such exchange paired with an improved documentation of research results would make scientific work more transparent and in the optimal case even reproducible. It would allow for the evaluation of fitness for further use (Devillers et al. Citation2007), and thus would ideally increase efficiency and sustainability of research investments and potentially stimulate interdisciplinary research. Simplified and refined representations of the data through Web-based visualizations and visual analysis tools for the comparison of different datasets could support stakeholder work and provide policy makers with insights from scientific research (Bernard and Ostländer Citation2008). If these various forms of reuse or at least the reuse of the data by other scientists is documented in a way comparable to literature citing, the evaluation of scientific outputs would get an additional measure.

The role of scientific information infrastructures in contributing to innovation and in addressing global challenges has been recognized and addressed in research funding initiatives like NSF (Citation2007) and ESFRI (Citation2008). Related Initiatives in America and Asia often use the label cyberinfrastructure whereas terms as e-science and e-Research Infrastructures have been attached to equivalent activities in Europe. We introduce the term Scientific GDI, as it is felt that current GDI provide a number of useful building blocks to support scientific collaboration and it is envisioned that Scientific GDI could become one of the core components in future Digital Earth. The US National Science Foundation specifically acknowledges the need for Scientific GDI by the Earth Cube program (http://www.nsf.gov/geo/earthcube) to foster research on so-called Geospatial Cyberinfrastructures (Yang et al. Citation2010). In general, scientific information infrastructures have been approached for different science domains and with differing functions, for example the social network myExperiment for the exchange of bioinformatics workflows (Goble et al. Citation2010) or the iPlant Collaborative for plant sciences (http://www.iplantcollaborative.org/). Central functions provided are access to data catalogues and the scientific data, data analysis and visualization capabilities, high performance computing and collaboration. Although such functionality can also be aided by GDI developments, the available scientific infrastructures for the exchange of data from the environmental and earth sciences hardly refer to corresponding interoperability standards. It is only recently that GDI concepts have been taken up by scientific initiatives. A prominent example is the launch of the data service of the Intergovernmental Panel on Climate Change (IPCC; http://www.ipcc-data.org/).

Creating a Scientific GDI () to support the environmental research community in efficiently exchanging data and methods related to the various scientific disciplines poses a number of diverse challenges on today's GDI developments. In revisiting the science case in Digital Earth (Craglia et al. Citation2012), one could consider an idealized and very simplified researchers workflow to structure these challenges:

  1. discover and access scientific geoinformation resources (data from observations, simulation and analysis as well as analysis methods and models),

  2. integrate, process and analyse scientific geoinformation resources,

  3. publish and share scientific geoinformation resources.

The remainder of the paper follows this sequence to address related issues and to hint to available approaches, best practices and potential solutions.

Figure 2. Aspects and core components of Scientific GDI.
Figure 2. Aspects and core components of Scientific GDI.

4. Challenges, approaches and recommendations for scientific GDI

4.1. Discover and access scientific geoinformation resources

Metadata captures the basic characteristics of a geoinformation resource and ideally does not only support information discovery (like in a library) but also allows assessing the fitness for use and enables the integration with other resources. In current GDI, metadata is usually acquired according to the structure and contents as defined in ISO 19115 and ISO 19119 (ISO Citation2003; ISO Citation2005b; ISO Citation2005a). These standards define more than 300 metadata elements. Therefore, the automatic acquisition of metadata (Manso-Callejo et al. Citation2004; Bockmühl, Mäs, and Reinhardt Citation2010) and further approaches towards user-based metadata enrichment (Olfat et al. Citation2012) are vital issues currently researched.

Typically, geoinformation metadata is managed in catalogues, which follow a quite static perspective on the generation of the underlying geoinformation resources (Fisher, Comber, and Wadsworth Citation2009). Further, standards lack in representing different levels of detail for geospatial metadata to address different (scientific) user communities.

A main requirement of publishing scientific data is the interlinking between the data and the corresponding scientific publications. This can be easily achieved by enriching metadata with references (cf. DOI in Section 4.3.1) that allows one to directly navigate to related publications. These publications can provide details about the scientific context, framework requirements of the research, inherent assumptions of the data processing and approaches of data interpretation. These aspects can hardly be fully covered by formalized metadata elements. Even assuming that future metadata sets do always provide these links and that these publications summarize information about the data producing methods, they are usually more focused on new scientific insights and, with regard to metadata provision, they do not sufficiently describe the data and its quality in a structured and comprehensible way. Therefore, links to publications are a part of but will not replace structured metadata.

INSPIRE, for example, requires that for a spatial dataset at least one keyword from the General Environmental Multilingual Thesaurus (GEMET) is provided to describe the relevant spatial data theme (EP Citation2007b). However, for the required scientific terms GEMET is insufficient. At present, there are other domain-specific vocabularies available or under development (e.g. WMO BUFR for atmospheric conditions (WMO Citation2010) or the planned GEOSS ontology). However, these are hardly used for unambiguously communicating the meaning across scientific disciplines. Some of these vocabularies reveal inconsistencies in their definitions when they are combined with others, such that vocabularies from different domains are partially incompatible. Beside the connection of keywords to common vocabularies, descriptions of data semantics are hardly considered in current geoinformation metadata standards, which are primarily concerned with the discovery of the data but do not yet enable data integration (Comber, Fisher, and Wadsworth Citation2008). Controlled vocabularies for generic scientific terms, like the science ontology developed in Brodaric (Citation2008) and Brodaric, Reitsma, and Qiang (Citation2008), hardly exist. Taking the example of environmental modelling, this already starts with basic concepts like model, storyline, scenario, driver and indicator. Although frequently used, different scientific communities have a slightly biased understanding of these terms. Creating a detailed and unambiguous formal description and, particularly, communicating it to a wider audience (at least within single scientific communities) is strongly required and a pressing challenge. Public fora as Wikipedia provide a blueprint on how a common corpus could be developed in a first step. However, additional organizational and research policy measures might be required to establish or even enforce the generation and usage of such common vocabularies.

In particular for climate change and economic development data, storylines and scenarios play a major role in metadata descriptions, as they can be used to classify the data and are certainly the most prominent keywords and search terms. The IPCC defines a storyline as ‘a narrative description of a scenario (or a family of scenarios), highlighting the main scenario characteristics and dynamics, and the relationships between key driving forces’ (Nakićenović, Alcamo, and Davis Citation2000). A storyline defines qualitative global constraints for projections and leaves a relatively big space for interpretations. Therefore, storylines are substantiated by scenarios, which are defined as ‘projections of a potential future’ (Nakićenović, Alcamo, and Davis Citation2000): Scenarios provide a ‘plausible description of how the future may develop, based on a coherent and internally consistent set of assumptions (scenario logic) about key relationships and driving forces (e.g. rate of technology change, prices).’ The IPCC defined a set of scenarios that are commonly used by scientists to obtain comparable model outputs when modelling effects of climate change. Therefore, metadata of these output data should contain unambiguous links to the scenarios and the corresponding storylines.

4.1.1. Accuracy and scale

Analysis and description of the quality and in particular the accuracy of the data are very often neglected by scientists. The elements for the description of accuracy and consistency of data defined by ISO 19115 (ISO Citation2003) are not fitting for scientific model outputs, because such descriptions are usually neither available nor very meaningful. In many cases it would be more helpful to get a description of relationships between the quality of input and output data and a rating or estimation of which of the inputs have the biggest impact on the quality of the model outputs. Also, there might be conditions described in the related metadata under which a value can be considered as accurate.

The current ISO metadata standards do not provide sufficient elements to adequately describe the spatio-temporal scale and level of detail of time series or multidimensional data. The emerging revision of ISO 19115 is already addressing this issue by allowing the statement of a temporal resolution. However, input datasets of numerical models very often include statistical data referencing discrete administrative units, like provinces and countries, as spatial resolution. We observed – for instance in economic models – that these units are not separately considered but aggregated to larger, equally sized regions to create a uniform sample size. Depending on modelling goals and the expected outputs, these aggregated spatial regions can be diverse and are task-specific. Nevertheless, the aggregation procedure is hardly documented once the data is published. Beside the spatial resolution, the scale of the geographical phenomena can also be diverse. Different objectives of models lead to different thematic categories in the data, like differing terminologies for land cover or agricultural products. For example, a generic class cereal crops could also be represented in more detail by wheat, rice, corn and barley, which can be considered as the most important subclasses in terms of agricultural production. For a cereal crops dataset, it should be documented if it refers to these four or also includes other ‘less important’ crops like, for example, millet. Thus, to support related up- and downscaling tasks, the metadata must contain resolvable links to the corresponding sets of spatial aggregation units and thematic categories, as for instance provided by the future INSPIRE implementations.

4.1.2. Lineage and usage

One of the main purposes of metadata within a GDI is to enable potential users to assess the suitability of the geodata for their specific use. In recent publications, available metadata standards have been recognized as data production oriented and it has been claimed for more user-centric metadata (Devillers et al. Citation2007; Goodchild Citation2007; Devillers et al. Citation2010). Due to its complexity and in particular the complexity of the data's provenance, this is especially true for scientific data. At present, even if scientific data is discoverable and accessible, the assessment of the data quality with regard to a particular use is difficult.

To evaluate the fitness for use of a dataset, information about its provenance is vital. From the data producer's perspective, lineage information can be used as an internal record of the data to ensure that the production standards are being maintained. In ISO 19115, such information is represented by the lineage element that is modelled as part of the data quality. Lineage recounts the life cycle of the dataset, from real-world abstraction, collection or acquisition, through all stages of compilations, corrections, maintenance, conversions and transformations to the generation of new interpreted products (Clarke and Clark Citation1995; Harding Citation2006; Servigne, Lesage, and Libourel Citation2006). The main components of the lineage in ISO 19115 are subclasses describing sources and process steps when generating the dataset. The sources provide information regarding the source data used to create the described dataset. Process steps are methods and processes used in the creation or maintenance of the dataset.

Since scientific environmental data is often an output of numerical models or simulations, the lineage sub-elements can be used to link to corresponding input datasets and models. If such links are systematically provided by metadata, the relationships between different datasets and models can be visually illustrated. The system MetaViz () has been realized to allow for using and visualizing provenance metadata in the GLUES GDI. Therewith, scientists can, for example, get a comprehensive view about which models provide datasets for a certain scenario or whether an input dataset also served into other models. Hitherto, gaining such information required a tedious, extensive and time consuming investigation of literature by each researcher who tried to learn about origination of a considered dataset.

Figure 3. MetaViz prototype showing an example for a world economy model. Access via http://geoportal.glues.geo.tu-dresden.de/geoportal/Applications/metaviz.html
Figure 3. MetaViz prototype showing an example for a world economy model. Access via http://geoportal.glues.geo.tu-dresden.de/geoportal/Applications/metaviz.html

Beside the scientific work, such comparison can also be of interest for research assessment, since it shows the ‘impacts’ of a dataset. This can also be useful when analysing the scientific outreach, since it represents the data exchange and collaboration between different research institutions. Therefore, lineage information can play a major role for the evaluation of scientific data, comparable to the way citations are used to rank scientific publications. Nevertheless, a detailed description of the origin of scientific data is not possible with the current ISO metadata. In particular the description lacks of details regarding the concrete model initialization, drivers and parameters and a generic description of the model and its basic assumptions.

4.2. Integrate, process and analyse scientific geoinformation resources

4.2.1. Software architectures

Current administrative GDI implementations are mostly realized as Service Oriented Architectures (SOA; (OASIS Citation2006; Erl Citation2007; OASIS Citation2012)). In contrast to open source programming libraries, SOA are following the black-box paradigm hiding the concrete functioning and implementation of an algorithm to the service client. Discovery, visualization, access and geoprocessing functions are made available as a set of distinct services. Each is following a well-defined interface to make the specific functions accessible and a service interoperable and easily replaceable by other services as well as combinable in different service orchestrations supporting a wide range of applications. Compared to previous generations of component based approaches or object broker architectures, SOA allow for a less tight coupling of the different service components but equally support realizing applications which require certain flexibility and complexity in the usage of processing functionality. Thus, SOA are for instance well suited when dealing with real-time observations, simulations and other data-intensive decision support applications, as for instance in environmental planning or in risk assessment and prevention.

Resource Oriented Architectures (ROA) follow the idea of simply considering anything in the Web a resource which could be easily linked with other resources. They gained widespread uptake in creating Mash Ups as simple information applications (e.g. maps on air quality or a cholera pandemic).

Linked Data follows the ROA path in envisioning a network of Web accessible datasets allowing mutual linkages and ideally referring to ontologies as semantic references (Kuhn Citation2005). SOA, ROA and Linked Data are addressing different levels of functional complexity and user requirements and would thus require research on their adequate layering (SOA-ROA-Linked Data) to achieve and enable their complimentary usage ().

Figure 4. Classification of Service Oriented Architectures, Resource Oriented Architectures and Linked Data.
Figure 4. Classification of Service Oriented Architectures, Resource Oriented Architectures and Linked Data.

Thus, ROA and Linked Data are proposed as the future path towards integration of SOA based GDI in non-geo Web domains and fostering their multidisciplinary usage. Metadata publication using Linked Data approaches can be recommended as the first step towards this goal. Clearly, not any observation taken every second or result sets of global change model simulation for 2100 would qualify as a Linked Data item (Janowicz et al. Citation2011), so that a combination of adequate aggregation and generalization mechanisms needs to be designed.

4.2.2. Sharing and integration of geoprocessing methods

Data processing logic in a GDI may be used for change and anomaly detection, to conduct general analysis as well as aggregation, fusion and transformation to convert, compare and integrate data that refers to different spatial, temporal and thematic granularities (ISO Citation2005b; Craglia et al. Citation2008; Brauner et al. Citation2009; Haubrock et al. Citation2009).

There are two fundamental strategies to deal with such data processing tasks in SDI (Müller, Bernard, and Brauner Citation2010). In data-driven approaches, geodata is sequentially shipped between a network of processing nodes, each offering a certain set of geoprocessing operators or other computational logic. Service chaining techniques as defined by ISO 19119 (ISO Citation2005b) are usually applied to create larger workflows by connecting data and processing services. Although a loose coupling of independent services is theoretically the most flexible approach, it has some major drawbacks in practical applications. First, purely data-driven workflows are also data-intensive and allocate a lot of network bandwidth. Second, it requires data owners to give their data to an unknown third party institution which prohibits to process license-, privacy- or security-constrained data in a federated GDI. Third, from a service provider perspective, offering facilities for computationally intensive jobs is a much greater commitment than the provision of a data download service.

Code-driven approaches pursue the opposite strategy – the code is shipped to the data nodes which is more challenging but offers better performance in terms of bandwidth efficiency, data privacy and provision of computational resources. Shipping code between different instances comes at the cost of the necessity to define exchange mechanisms for algorithms as the definition of common algebras and processing languages. Some of the OGC web services can be extended with basic processing capabilities by supporting standardized processing languages: The Filter Encoding specification (OGC Citation2010b) is supported by a variety of Web Feature Services (WFS, (ISO Citation2010)); for Web Coverage Services (WCS, (OGC Citation2010a)) the Web Coverage Processing Service (WCPS) Language Interface Standard (OGC Citation2009) plays a similar role. Both standards follow the principle of database query languages that allow their users to perform simple sub-setting and arithmetic tasks on the contained data. Instead of shipping the data to a processing service that performs the intended computations, the client may send a processing query directly to an extended data service where the statements are interpreted and processed. This circumvents the necessity to chain independent data and processing services to accomplish a processing task. A detailed comparison of both approaches was carried out in the OWS-8 testbed (OGC Citation2011).

Moving and sharing computational code as freely as data in a GDI would be the most flexible solution for code-driven scenarios and might operate equally well as data-driven strategies without suffering from the performance drawbacks. A more general framework for code exchange requires an agreement on common functionality and computing platforms, including software environments and infrastructure. Some research has been done on that matter (Müller, Bernard, and Brauner Citation2010; Müller, Kadner, and Bernard Citation2012) which demonstrates the feasibility of code-driven strategies in a GDI but has not yet led to any operational standard or widely used practice. However, the concept was found to link nicely with cloud computing since portability of the software stack is a core asset to achieve computing scalability.

Considering the efforts by individual institutions to foster software reuse, sharing software in larger groups seems to be the next logical step. Code sharing approaches may spur this development and increase the availability of reusable and portable analysis tools, which have been identified as a cornerstone in data-driven science (Gray Citation2009).

4.2.3. Sharing and integration of environmental simulation models

Research and developments on linking environmental models from different disciplines, as for example coupling a hydrological model with a climate model, and linking environmental models with GIS to facilitate the pre-processing of model inputs and the analysis of the spatio-temporal results has a long tradition in environmental and geoinformation science (Goodchild, Parks, and Steyaert Citation1993; Goodchild, Steyaert, and Parks Citation1996; Bernard and Krüger Citation2000; Haubrock et al. Citation2009; Maué et al. Citation2011). Roughly three ways of sharing and integrating of environmental models can be distinguished:

  1. Collective model development matured in a number of community models. Very equally to open source projects these communities from one or several research institutions normally focus on the realization of a common model for a certain discipline (e.g. the Weather Research & Forecasting Model, http://www.wrf-model.org).

  2. Sharing model components with the goal to allow an integrated usage of models from different disciplines is tackled by different approaches: Existing models are coupled through the usage of common interfaces (e.g. following the IEEE standards on High Level Architecture (IEEE Citation2010) or the Open Modelling Interface http://www.openmi.org/). Integrated models get (newly) developed to synthesize existing models from various disciplines and following different modelling styles (Beven Citation2007; Voinov Citation2010).

  3. Sharing simulation results allowing scientists from other disciplines to access the results from specific model simulations and scenarios. Results are either stemming from a previous simulation (as for the IPCC data) and should have undergone the required plausibility checks and validations, or access is given to a pre-parameterized, robust model, which allows for an ad-hoc execution (e.g. an interpolation model to provide a spatial distribution on air pollution, (Stasch et al. Citation2012; Wiemann et al. Citation2012)). In considering simulation models as sensors, the suite of Sensor Web Enablement interfaces provide means to share simulation results in GDI (Maué et al. Citation2011).

These sharing paradigms can be seen as different sharing levels, which could be stacked according to the required need of expertise and the expected users and usages ().

Figure 5. Layers in sharing and integrating environmental models.
Figure 5. Layers in sharing and integrating environmental models.

GIS and GDI noticeably improved in their capabilities to manage and visualize spatio-temporal (3D) geodata, however, related processing methods (e.g. to calculate a spatio-temporal mean or sum when dealing with time-variant 3D atmospheric data) are still mostly lacking. Advancing these tools and thus better supporting data-mining on environmental simulation results and observations could provide a better means to design integrated models. These integrated models are expected to act as the glue between the more complex and demanding models within the environmental disciplines and would ideally become available in a Scientific GDI as robust models to serve tools for other scientific communities and decision makers. As stated above for metadata, handling and communicating the assumptions and uncertainties underlying a model are pressing demands which should be tackled before further encouraging the sharing of environmental models and simulation results. Approaches to consider related metadata in the further processing and analysis chain – as proposed by Pebesma, de Jong, and Briggs (Citation2007); Jung-Hong and Min-Lang (Citation2012) – should be further elaborated and find their way in operational GIS and GDI implementations.

4.3. Publish and share scientific geoinformation resources

4.3.1. Publishing scientific data, methods and models

Although some scientific journals offer possibilities to link publications with data or software sources, there are not many incentives for researchers to publish their data or software. Scientific impact factors only consider textual publications and their citations. To better stimulate the publication of scientific data and scientific software, they must be properly acknowledged, e.g. by having impact factors for data quotation and by assuring intellectual property rights for provided scientific data and methods. This links to current activities to make publications of scientific data and models obligatory for research work which gains public funding (EC Citation2012).

Published scientific data should be persistently identifiable and available in a way comparable to current textual publications. A recent open letter of the US National Science Foundation GEO Directorate exemplifies this issue and strongly recommends data citation (http://www.nsf.gov/pubs/2012/nsf12058/nsf12058.jsp). The digital object identifier (DOI) provides a system for identifying content objects in the digital environment. Data DOIs provide persistent identifiers and long-term access to data. They help researchers to find and cite datasets, for example to cross reference journal articles with the underlying data. DOI is an ISO standard (ISO Citation2005a) and is already used by numerous platforms for referencing data publications, for example PANGAEA (Data Publisher for Earth and Environmental Science; http://www.pangaea.de/) and the World Data Center for Climate in Hamburg (http://cera-www.dkrz.de/CERA/). Data DOI registration services are provided by the Data Cite organization (http://datacite.org/) which operates globally with national representations. Existing GDI hardly make use of DOI as unique data references. Conversely, data that has a unique DOI reference does not necessarily have standardized metadata or use standardized data formats or geoinformation services. Making DOI a requirement when scientific data is published in a GDI seems to be indispensable. Nevertheless, the systematic capturing of DOI citations and corresponding rewarding of the publishers is still an issue.

The unique identification of data can be implemented in different levels of granularity. INSPIRE, for example, defines persistent object IDs for individual instances. There might be cases which require that single objects be referenced in such a way. However, the overall efforts to establish and maintain such detailed identification are considered to be disproportionate to the benefits in most Scientific GDI. Thus, identification on the coarser grained dataset level seems recommendable for using geodata DOIs in Scientific GDI.

4.3.2. Licensing

A major concern of scientists publishing their data is about licensing, intellectual property rights and warranty. Possible users must be informed about the concrete license of available data. The Creative Commons (CC) licensing framework provides a set of predefined harmonized and simplified licenses that can be combined to a license contract. CC licenses are commonly used in scientific data infrastructures. Usually, the data ownership stays with the data producing institution or scientist. The expectations of the providers of the scientific data and the infrastructure regarding citation, credits or acknowledgement must be clearly communicated and follow the ‘good scientific practices’. The data producers should be aware that the selection of restrictive license conditions possibly discourages potential users, for example when a license prohibits commercial use. Recently, the CC community is seeking to adjust its licenses to the specific requirements of scientific data (http://creativecommons.org/science). Scientific GDI must also guarantee the confidentiality of published data to assure the competition between scientific teams. A common strategy is, for instance, to make only the metadata available and to provide the actual data access later (e.g. after two years), when the data producers have finished corresponding publications (e.g. HALO project database, http://halo-db.tropos.de/).

Another issue is the access to privacy restricted data. Common privacy rules are relatively strict and in some cases they prohibit the provision of data required to explore or analyse detailed relations or interdependencies (e.g. in health or socio-economic studies). In general, GDI still lack the possibility to formalize license and privacy policies, to allow for an automated policy negotiation and enforcement process for research applications.

4.3.3. Governance and policies

Most of the current GDI initiatives and implementations are driven by public administrations and result from legal enforcement, like the European INSPIRE initiative. In contrast, commercial online mapping applications successfully followed a kind of ‘supply creates demand’ approach (Craglia et al. Citation2012). A mixture of both might stimulate Scientific GDI. Legal measures will hardly work for Scientific GDI, steering needs to relate to the various research funding mechanisms (e.g. simply by making the reuse of existing data and the publishing of results an unescapable requirement for research funding) and rewarding cultures (citation indices, best data awards, etc.). Having specific Scientific GDI implementations in place, which are convincing in terms of their data richness, usability and acceptance by a scientific community, further demand can be generated and stimulate further Scientific GDI activities. Initiatives as the CUAHSI Hydrologic Information System (http://his.cuahsi.org), PANGAEA (http://www.pangaea.de/), the EuroGEOSS broker (http://www.eurogeoss-broker.eu/) or the GLUES GDI on sustainable land management and ecosystem research (Eppink et al. Citation2012) could for instance serve as such stimuli on Scientific GDI.

Assuring the sustainable operation and availability of future nodes in the Scientific GDI is another challenge. In general, scientific infrastructures should offer a reliable service with long-term perspective and aim at a long-term preservation of digital assets. Since they should be sustained beyond the lifetime of single research projects or programmes, the funding must also be sustainable and independent of projects or research programmes and most likely linked to suited research institutions, which would act as the backbone and service centre in Scientific GDI. Current INSPIRE experiences could help in identifying the best suited organizational models. As the INSPIRE directive also enforces specific service level agreements, it could provide guidance for defining comparable criteria in implementing the central nodes of a Scientific GDI. Consequently, Scientific GDI are supposed to follow a system of systems pattern, where such reliable central nodes would serve a wider scientific community in discovering the more specific scientific infrastructures, which could be bound to a specific region or discipline.

5. Conclusion

Similarly to what De Vries et al. (Citation2011) observe about INSPIRE implementations of national GDIs, best practices for the implementation of Scientific GDIs have been identified, but the choice for a certain practice and prioritizing the issues always bounds to the concrete contextual conditions and opportunities. Having broadly considered issues on the realization of Scientific GDI and based on our various experiences in realizing Scientific GDI, we conclude in prioritizing related research items for a Scientific GDI 2020 agenda. Again, as for Section 4, using the three areas (a) discovery and access, (b) integration, processing and analysis, and (3) publish and share of scientific geoinformation resources, we propose the following major research activities for this areas:

  1. Discovery and access

    1. (Further) establish and possibly enhance common standards and reference systems (e.g. common vocabularies) to improve interoperability especially in terms of semantics.

    2. Improve methods for automated metadata creation, extraction, maintenance and provision to allow for better assessment and integration of the provided resources also addressing descriptions on provenance, accuracy, scale and models.

    3. Advance usability of existing scientific data portals and services, to attract more users and usage, both in terms of data access and data provision.

    4. Provide technical concepts to master the balancing act between scientific data products for scientists and mass-market users.

  2. Integration, processing and analysis

    1. Allow for efficiently exchanging analysis methods (processing logic) to better support interdisciplinary and distributed data processing.

    2. Further advance methods in multi-scale transformations and cross-disciplinary data integration.

    3. Enhance GDI in dealing with time and spatio-temporal information.

    4. Progress on having commonly accepted means to exchange and integrate environmental models and the related simulation results within and across the several thematic domains by further aligning the existing partly different developments towards community models, open modelling interfaces and sensor Web.

    5. Explore ways to link traditional sensor measurements with crowd sourced observations.

  3. Publish and share

    1. Further establish common policies and organizational frameworks for (cross-disciplinary) scientific data infrastructures within the frame of research infrastructures (e.g. http://ec.europa.eu/research/infrastructures), where Scientific GDI would become an integral part.

    2. Better stimulate scientific data and scientific software publications by properly acknowledging these kind of publications, e.g. by having impact factors for data quotation and by assuring intellectual property rights for provided scientific data and methods.

    3. Progress in formalizing license and privacy policies for research applications to allow for an as much as possible automated policy negotiation and enforcement, as well as to ease analysis of privacy restricted data (e.g. in health or socio-economic studies).

Some of the issues addressed in this agenda have not been covered within this paper but would deserve treatment and further submissions in their own rights. Improving usability for instance is regarded to be of utmost importance (1c and 1d in the list above). Balancing functionality and simplicity is a prerequisite to achieve convincing usability and broadly accepted information infrastructures. However, current GDI and GIS developments are still felt very weak in terms of usability. Another example is formalizing license and privacy policies (3c) as a foundation to allow for distributed geoprocessing. Today, various databases are not accessible for scientific analyses or can only be used in an off-line mode to ensure data and privacy protection. Thus, for instance, research on health-environment effects or socio-economic patterns is not only tedious but partly even hampered as links between different data sources cannot be generated. Beside a need for harmonized policies and organizational frameworks to best enable scientific analysis on such databases, it also lacks commonly agreed formalisms and technical mechanisms. In the end, these technologies should allow for an automated policy negotiation and enforcement and an on-line usage of the restricted database for authorized scientists.

Clearly, the research agenda proposed here closely links to current GEOSS and Digital Earth activities. These are expected to provide the global and cross-disciplinary frame for the emerging more regional or discipline-bounded Scientific GDI advancements.

Acknowledgements

Several findings presented in this paper stem from work on the GLUES GDI as part of the research programme ‘Sustainable Land Management’ funded by the German Federal Ministry of Education and Research (BMBF). The comments and suggestions of two anonymous reviewers and the journal editors helped to improve this submission and are gratefully acknowledged.

References

  • Bernard, L., I. Kanellopoulos, A. Annoni, and P. Smits. 2005. “The European Geoportal – One step towards the Establishment of a European Spatial Data Infrastructure.” Computers, Environment and Urban Systems 29 (1): 15–31. http://www.sciencedirect.com/science/article/pii/S0198971504000493
  • Bernard, L., and T. Krüger. 2000. “Integration of GIS and Spatio-temporal Simulation Models.” Transactions in GIS 4 (3): 197–215. doi:10.1111/1467-9671.00049
  • Bernard, L., and N. Ostländer. 2008. “Assessing Climate Change Vulnerability in the Arctic Using Geographic Information Services in Spatial Data Infrastructures.” Climatic Change 87 (12): 263–281. http://www.springerlink.com/content/a786301673025334/fulltext.pdf
  • Beven, K. 2007. “Towards Integrated Environmental Models of Everywhere: Uncertainty, Data and Modelling as a Learning Process.” Hydrology and Earth System Sciences 11 (1): 460–467. http://www.hydrol-earth-syst-sci.net/11/460/2007/
  • Bockmühl, T., S. Mäs, and W. Reinhardt. 2010. “An Approach for Automatic Generation of Metadata based on Data Product Specifications”. INSPIRE Conference 2010. Krakow, Poland: JRC, European Commision. Accessed 20 May 2012. http://inspire.jrc.ec.europa.eu/events/conferences/inspire_2010/abstracts/143.doc
  • Brauner, J., T. Foerster, B. Schaeffer, and B. Baranski. 2009. “Towards a Research Agenda for Geoprocessing Services.” In Proceedings of the 12th AGILE Conference, edited by M. Seester, L. Bernard, and V. Paehlke. Hannover: AGILE. Accessed 20 May 2012. http://agile.gis.geo.tu-dresden.de/web/Conference_Paper/CDs/AGILE%202009/AGILE_CD/pdfs/124.pdf
  • Brodaric, B., 2008. Science Knowledge Infrastructure Ontology 3.0. University of Edinburgh. Accessed 20 May 2012. http://www.nesc.ac.uk/technical_papers/UKeS-2008-03.pdf
  • Brodaric, B., F. Reitsma, and Y. Qiang. 2008. “SKIing with DOLCE: Toward an e-Science Knowledge Infrastructure.” In Fifth International Conference on Formal Ontology in Information Systems (FOIS 2008), edited by C. Eschenbach and M. Grüninger, 208–219. Amsterdam, The Netherlands: IOS Press.
  • Clarke, D. G., and D. M. Clark. 1995. “Lineage.” In Elements of Spatial Data Quality, edited by S. C. Guptill and J. L. Morrison, 13–30. Oxford, UK: Oxford Elsevier Science.
  • Comber, A. J., P. F. Fisher, and R. A. Wadsworth. 2008. “Semantics, Metadata, Geographical Information and Users.” Transactions in GIS 12 (3): 287–291. doi:10.1111/j.1467-9671.2008.01102.x
  • Craglia, M., K. de Bie, D. Jackson, M. Pesaresi, G. Remetey-Fülöpp, C. Wang, A. Annoni, et al. 2012. “Digital Earth 2020: Towards the Vision for the Next Decade.” International Journal of Digital Earth 5 (1): 4–21. doi:10.1080/17538947.2011.638500
  • Craglia, M., M. F. Goodchild, A. Annoni, G. Camara, M. Gould, W. Kuhn, D. Mark, et al. 2008. “Next-generation Digital Earth: A Position Paper from the Vespucci Initiative for the Advancement of Geographic Information Science.” International Journal of Spatial Data Infrastructure Research 3, 146–167. Accessed 20 May 2012. http://ijsdir.jrc.ec.europa.eu/index.php/ijsdir/article/download/119/99.
  • Devillers, R., Y. Bedard, R. Jeansoulin, and B. Moulin. 2007. “Towards Spatial Data Quality Information Analysis Tools for Experts Assessing the Fitness for Use of Spatial Data.” International Journal of Geographical Information Science 21 (3): 261–282. doi:10.1080/13658810600911879
  • Devillers, R., A. Stein, Y. Bédard, N. Chrisman, P. Fisher, and W. Shi. 2010. “30 Years of Research on Spatial Data Quality: Achievements, Failures, and Opportunities.” Transactions in GIS 14 (4): 387–400. doi:10.1111/j.1467-9671.2010.01212.x
  • De Vries, W., J. Crompvoets, J. Stoter, and I. Vanden Berghe. 2011. “ATLAS of INSPIRE – Conceptualizing SDI Implementation through an Inventory of Experiences, Successes and Headaches of European National Mapping Agencies.” International Journal of Spatial Data Infrastructures Research 6, 97–125. Accessed 20 May 2012. http://ijsdir.jrc.ec.europa.eu/index.php/ijsdir/article/download/208/290
  • EC. 2011. MEMO/11/891: Digital Agenda: Commission's Open Data Strategy, Questions & Answers [online]. European Commission. Accessed 20 May 2012. http://europa.eu/rapid/pressReleasesAction.do?reference=MEMO/11/891
  • EC. 2012. Scientific Data: Open Access to Research Results will Boost Europe's Innovation Capacity. European Commission. Accessed 20 May 2012. http://europa.eu/rapid/pressReleasesAction.do?reference=IP/12/790
  • EP. 2007a. Directive 2007/2/EC of the European Parliament and of the Council of 14 March 2007 establishing an Infrastructure for Spatial Information in the European Community (INSPIRE)
  • EP. 2007b. Implementing Directive 2007/2/EC of the European Parliament and of the Council as regards metadata (Commission Regulation (EC) No 1205/2008 of 3 December 2008).
  • Eppink, F., A. Werntze, S. Mäs, A. Popp, and R. Seppelt. 2012. “Land Management and Ecosystem Services: How Collaborative Research Programmes Can Support Better Policies.” GAIA - Ecological Perspectives for Science and Society 21 (1): 55–63. Accessed 20 May 2012. http://www.ingentaconnect.com/content/oekom/gaia/2012/00000021/00000001/art00014
  • Erl, T. 2007. SOA Principles of Service Design. Boston, US: Prentice Hall International.
  • ESFRI. 2008. European Roadmap for Research Infrastructures [Online]. European Strategy Forum on Research Infrastructures. Accessed June 11, 2012. ftp://ftp.cordis.europa.eu/pub/esfri/docs/esfri_roadmap_update_2008.pdf
  • Fisher, P., A. J. Comber, and R. Wadsworth. 2009. “What's in a Name? Semantics, Standards and Data Quality.”In Spatial Data Quality: Form Process to Decisions, edited by R. Devillers and H. Goodchild, 3–16. Boca Raton, US: CRC Press.
  • Goble, C. A., J. Bhagat, S. Aleksejevs, D. Cruickshank, D. Michaelides, D. Newman, M. Borkum, et al. 2010. “myExperiment: A Repository and Social Network for the Sharing of Bioinformatics Workflows.” Nucleic Acids Research 38, W677–W682. doi:10.1093/nar/gkq429
  • Goodchild, M. F. 2007. Beyond Metadata: Towards User-Centric Description of Data Quality. 5th International Symposium on Spatial Data Quality. Enschede, The Netherlands: ITC.
  • Goodchild, M. F. 2008. “The Use Cases of Digital Earth.” International Journal of Digital Earth 1 (1): 31–42. doi:10.1080/17538940701782528
  • Goodchild, M. F., and J. A. Glennon. 2010. “Crowdsourcing Geographic Information for Disaster Response: A Research Frontier.” International Journal of Digital Earth 3 (3): 231–241. doi:10.1080/17538941003759255
  • Goodchild, M. F., B. O. Parks, and L. T. Steyaert, eds. 1993. Environmental Modeling with GIS. New York: Oxford University Press.
  • Goodchild, M. F., L. T. Steyaert, and B. O. Parks 1996. GIS and Environmental Modeling: Progress and Research Issues. Fort Collins: GIS World Books.
  • Gore, A. 1999. “The Digital Earth: Understanding Our Planet in the 2lst Century (Speech held on January 31, 1998 in the California Science Center).” Photogrammetric Engineering & Remote Sensing 65 (5): 528–530. Accessed 20 May 2012. http://www.asprs.org/a/publications/pers/99journal/may/highlight2.pdf
  • Gray, J. 2009. “eScience: A Transformed Scientific Method.” In The fourth Paradigm: Data-intensive Scientific Discovery, edited by A. J. G. Hey, S. Tansley, and K. M. Tolle, xvii–xxxi. Redmond, WA: Microsoft Research.
  • Harding, J. D. 2006. “Vector Data Quality: A Data Provider's Perspective.” In Fundamentals of Spatial Data Quality, edited by R. Devillers and R. Jeansoulin, 141–159. London: ISTE.
  • Haubrock, S., F. Theisselmann, H. Rotzoll, and D. Dransch. 2009. “Web-based Management of Simulation Models – Concepts, Technologies and the Users' Needs.” In 18th World IMACS Congress and MODSIM09 International Congress on Modelling and Simulation, edited by R. S. Anderssen, R. D. Braddock, and L. T. H. Newham, 880–886. Cairns, Australia: Modelling and Simulation Society of Australia and New Zealand and International Association for Mathematics and Computers in Simulation.
  • IEEE. 2010. 1516 IEEE Standard for Modeling and Simulation (M&S) High Level Architecture (HLA) Framework and Rules. New York: Institute of Electrical and Electronics Engineers.
  • ISO 19115 International Standard on Geographic information – Metadata 2003. Oslo, Norway: ISO/TC 211 Secretariat, Standards Norway.
  • ISO 19115–2 International Standard on Geographic information – Part 2: metadata for imagery and gridded data 2005a. Oslo, Norway: ISO/TC 211 Secretariat, Standards Norway.
  • ISO 19119 International Standard Geogaphic information Services 2005b. Oslo, Norway: ISO/TC 211 Secretariat, Standards Norway.
  • ISO 19142 Geographic information - Web Feature Service 2010. Oslo, Norway: ISO/TC 211 Secretariat, Standards Norway.
  • Janowicz, K., A. Bröring, C. Stasch, S. Schade, T. Everding, and A. Llaves. 2011. “A RESTful Proxy and Data Model for Linked Sensor Data.” International Journal of Digital Earth, in press. doi:10.1080/17538947.2011.614698.
  • Jung-Hong, H., and H. Min-Lang. 2012. “Interoperable GIS Operations: A Quality-aware Perspective.” In 15th AGILE International Conference on Geographic Information Science, edited by J. Gensel, D. Josselin, and D. Vandenbroucke. Avignon, France: AGILE. Accessed 20 May 2012. http://agile.gis.geo.tu-dresden.de/web/Conference_Paper/CDs/AGILE%202012/proceedings/papers/Paper_Hong_Interoperable_GIS_Operations_A_Quality-Aware_Perspective_2012.pdf
  • Kuhn, W. 2005. “Geospatial Semantics: Why, of What, and How?” In Journal on Data Semantics III, edited by S. Spaccapietra and E. Zimányi, 1–24. Berlin: Springer.
  • Manso-Callejo, M. A., J. Nogueras-Iso, J. Zarazaga, and M. A. Bernabe-Poveda. 2004. “Automatic Metadata Extraction from Geographic Information.” In 7th AGILE Conference on Geographic Information Science, edited by F. Toppen, and P. Prastacos, 379–386. Heraklion, Greece: AGILE.
  • Masser, I. 1999. “All Shapes and Sizes: The First Generation of National Spatial Data Infrastructures.” International Journal of Geographical Information Science 13 (1): 67–84. doi:10.1080/136588199241463
  • Maué, P., C. Stasch, G. Athanasopoulos, and L. Gerharz. 2011. “Geospatial Standards for Web-enabled Environmental Models.” International Journal of Spatial Data Infrastructures Research 6: 145–167. http://ijsdir.jrc.ec.europa.eu/index.php/ijsdir/article/view/198/291
  • Müller, M., L. Bernard, and J. Brauner. 2010. “Moving Code in Spatial Data Infrastructures – Web Service Based Deployment of Geoprocessing Algorithms.” Transactions in GIS 14: 101–118. http://onlinelibrary.wiley.com/doi/10.1111/j.1467-9671.2010.01205.x/pdf
  • Müller, M., D. Kadner, and L. Bernard. 2012. “Moving Code - Sharing Geospatial Computation Logic on the Web.” In 15th AGILE International Conference on Geographic Information Science, edited by J. Gensel, D. Josselin, and D. Vandenbroucke, 58–62. Avignon: AGILE. Accessed 20 May 2012. http://agile.gis.geo.tu-dresden.de/web/Conference_Paper/CDs/AGILE%202012/proceedings/papers/Paper_Mueller_Moving_Code-Sharing_Geospatial_Computation_Logic_on_the_Web_2012.pdf
  • Nakićenović, N., J. Alcamo, and G. Davis. 2000. IPCC Special Report on Emissions Scenarios (SRES). Cambridge, United Kingdom: Cambridge University Press.
  • NSF. 2007. Cyberinfrastructure Vision for 21st Century Discovery [online]. US National Science Foundation Cyberinfrastructure Council. Accessed 20 May 2012. http://www.nsf.gov/od/oci/CI_Vision_March07.pdf
  • OASIS. 2006. Reference Model for Service Oriented Architecture 1.0 OASIS Standard, 12 October 2006. OASIS. Accessed 20 May 2012. http://docs.oasis-open.org/soa-rm/v1.0/soa-rm.pdf
  • OASIS. 2012. Reference Architecture Foundation for Service Oriented Architecture Version 1.0. 04 December 2012. OASIS Committee Specification Draft 03. OASIS. Accessed 20 May 2012. http://docs.oasis-open.org/soa-rm/soa-ra/v1.0/soa-ra.pdf
  • OGC. 2009. OpenGIS Web Coverage Processing Service (WCPS) Language Interface Standard [online]. Open Geospatial Consortium. http://www.opengeospatial.org/standards/wcps
  • OGC. 2010a. OGC WCS 2.0 Interface Standard - Core. Open Geospatial Consortium. http://www.opengeospatial.org/standards/wcs/
  • OGC. 2010b. OpenGIS Filter Encoding 2.0 Encoding Standard [online]. Open Geospatial Consortium. http://www.opengeospatial.org/standards/filter
  • OGC. 2011. OWS-8 Geoprocessing for Earth Observations Engineering Report [online]. Open Geospatial Consortium. http://www.opengeospatial.org/pub/www/ows8/deliverables.html
  • Olfat, H., M. Kalantari, A. Rajabifard, H. Senot, and I. P. Williamson. 2012. “Spatial Metadata Automation: A Key to Spatially Enabling Platform.” International Journal of Spatial Data Infrastructures Research 7, 173–195. Accessed 20 May 2012. http://ijsdir.jrc.ec.europa.eu/index.php/ijsdir/article/download/279/326.
  • Pebesma, E. J., K. de Jong, and D. Briggs. 2007. “Interactive Visualization of Uncertain Spatial and Spatio-Temporal Data Under Different Scenarios: An Air Quality Example.” International Journal of Geographical Information Science 21 (5): 515–527. http://dx.doi.org/10.1080/13658810601064009
  • Percivall, G. 2010. “The Application of Open Standards to Enhance the Interoperability of Geoscience Information.” International Journal of Digital Earth 3 (S1): 14–30. doi:10.1080/17538941003792751
  • Schäffer, B., B. Baranski, and T. Foerster. 2010. “Towards Spatial Data Infrastructures in the Clouds.” In Geospatial Thinking - 13th AGILE International Conference on Geographic Information Science, edited by M. Painho, M. Santos, and H. Pundt, 399–418. Berlin: Springer.
  • Servigne, S., N. Lesage, and T. Libourel. 2006. “Quality Components, Standards, and Metadata.” In Fundamentals of Spatial Data Quality, edited by R. Devillers and R. Jeansoulin, 179–210. London: ISTE.
  • Stasch, C., T. Foerster, C. Autermann, and E. Pebesma. 2012. “Spatio-temporal Aggregation of European Air Quality Observations in the Sensor Web.” Computers & Geosciences, 47: 111–118. doi:10.1016/j.cageo.2011.11.008.
  • Voinov, A. A. 2010. “‘Integronsters’ and the Special Role of Data.” In iEMSs Fourth Biennial Meeting: International Congress on Environmental Modelling and Software (iEMSs 2010). International Environmental Modelling and Software Society, edited by D. A. Swayne, W. Yang, A. A. Voinov, A. Rizzoli, and T. Filatova, Ottawa, Ontario, S.10.19. Accessed 20 May 2012. http://www.iemss.org/iemss2010/papers/S10/S.10.19.Integronsters%20and%20the%20special%20role%20of%20data%20-%20ALEXEY%20VOINOV.pdf
  • Wiemann, S., and L. Bernard. 2010. “Conflation Services within Spatial Data Infrastructures.” In 13th AGILE International Conference on Geographic Information Science, edited by M. Painho, M. Y. Santos, and H. Pundt, 1–8. Guimarães, Portugal: AGILE. Accessed 20 May 2012. http://agile.gis.geo.tu-dresden.de/web/Conference_Paper/CDs/AGILE%202010/ShortPapers_PDF/134_DOC.pdf
  • Wiemann, S. S. R., P. Karrasch, J. Brauner, K. Pech, and L. Bernard. 2012. “Classification-Driven Air Pollution Mapping as for Environment and Health Analysis.” In iEMSs Sixth Biennial Meeting: International Congress on Environmental Modelling and Software (iEMSs 2012). International Environmental Modelling and Software Society, edited by R. Seppelt, A. A. Voinov, S. Lange, and D. Bankamp. Halifax, Canada: International Environmental Modelling and Software Society (iEMSs). Accessed 20 May 2012. http://www.iemss.org/sites/iemss2012//proceedings/B1_0827_Wiemann_et_al.pdf
  • WMO. 2010. Manual on Codes (WMO-No. 306). World Meteorological Organization. Accessed 20 May 2012. http://www.wmo.int/pages/prog/www/WMOCodes.html
  • Yang, C., M. Goodchild, Q. Huang, D. Nebert, R. Raskin, Y. Xu, M. Bambacus, and D. Fay. 2011. “Spatial Cloud Computing: How can the Geospatial Sciences Use and Help Shape Cloud Computing?.” International Journal of Digital Earth 4 (4): 305–329. doi:10.1080/17538947.2011.587547
  • Yang, C., R. Raskin, M. Goodchild, and M. Gahegan. 2010. “Geospatial Cyberinfrastructure: Past, Present and Future.” Computers, Environment and Urban Systems 34 (4): 264–277. doi:10.1016/j.compenvurbsys.2010.04.001

Reprints and Corporate Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

To request a reprint or corporate permissions for this article, please click on the relevant link below:

Academic Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

Obtain permissions instantly via Rightslink by clicking on the button below:

If you are unable to obtain permissions via Rightslink, please complete and submit this Permissions form. For more information, please visit our Permissions help page.