6,693
Views
56
CrossRef citations to date
0
Altmetric
Review Article

A survey of methods incorporating spatial information in image classification and spectral unmixing

, , , &
Pages 3870-3910 | Received 03 Mar 2016, Accepted 17 May 2016, Published online: 13 Jul 2016

ABSTRACT

Over the past decade, the incorporation of spatial information has drawn increasing attention in multispectral and hyperspectral data analysis. In particular, the property of spatial autocorrelation among pixels has shown great potential for improving understanding of remotely sensed imagery. In this paper, we provide a comprehensive review of the state-of-the-art techniques in incorporating spatial information in image classification and spectral unmixing. For image classification, spatial information is accounted for in the stages of pre-classification, sample selection, classifiers, post-classification, and accuracy assessment. With regards to spectral unmixing, spatial information is discussed in the context of endmember extraction, selection of endmember combinations, and abundance estimation. Finally, a perspective on future research directions for advancing spatial-spectral methods is offered.

1. Introduction

Image classification and spectral unmixing are two predominant methods used for multispectral and hyperspectral data analysis. Along with the advancement of both satellite remote sensing and spatial science, spatial information has attracted significant interest within the remote-sensing community. Numerous studies have been conducted with the aim of effectively incorporating spatial autocorrelation in both image classification and spectral unmixing. However, compared to spectral information, there are few summaries of recent efforts to investigate how spatial information is manifested and their effectiveness in remote-sensing analysis. In this paper, we provide a comprehensive review of state-of-the-art techniques used to incorporate spatial information in image classification and spectral unmixing.

2. Incorporating spatial information in image classification

In this section, we summarize the state-of-the-art methods that incorporate spatial information in the image classification process according to the following five categories: (1) pre-classification, (2) sample selection, (3) classifiers, (4) post-classification, and (5) accuracy assessment.

2.1. Pre-classification

2.1.1. Image texture

2.1.1.1. Grey-level co-occurrence matrix (GLCM)

The grey-level co-occurrence matrix (GLCM), introduced by Haralick, Shanmugam, and Dinstein (Citation1973), is probably the most commonly adopted texture extraction method for remote-sensing images. A GLCM records the frequency of pairs of pixel values, where the pairs are oriented in a particular direction (0°, 45°, 90°, 135°), that occur in an image. Various texture measures, such as the mean, variance, homogeneity, contrast, dissimilarity, entropy, second moment, and correlation, can then be calculated from this matrix. The GLCM (and thus the texture measures) are usually calculated within a moving window to generate the texture images. While many studies have explored different GLCM texture measures and the role of different window sizes to better characterize image texture, very few have attempted to improve the statistical method itself. Clausi and Zhao (Citation2002) used a hash table and linked list (grey-level co-occurrence hybrid structure (GLCHS)) to store the co-occurring probabilities, resulting in a significant reduction in the computation time for calculating the co-occurrence texture features. Later, the GLCHS algorithm was again improved using a grey-level co-occurrence hybrid histogram (GLCHH) for sum and difference histograms (Clausi and Zhao Citation2003). By integrating GLCHS and GLCIA, the resulting grey-level co-occurrence integrated algorithm (GLCIA) requires only 0.04–16% of the computational time of the standard GLCM method to determine co-occurrence probability texture features (837–850). More recently, Huang, Liu, and Zhang (Citation2014a) applied clustering and sparse representation to multispectral images to generate codes or quantized levels based on the multi-channel texture features to be extracted. As a result, the traditional single-band GLCM method was improved so that multi-/hyperspectral imagery features can be extracted. Similarly, Tsai and Lai (Citation2013) were able to extract texture features from image cubes rather than 2-D neighbourhoods by extending the original co-occurrence matrix to a 3-D tensor field.

2.1.1.2. Geostatistical approach

Geostatistical methods have been widely applied to characterize the spatial relationships among pixel values of a particular remote-sensing image (i.e. texture) (Atkinson and Lewis Citation2000). The variogram is a summary of how semivariance (a measure of autocorrelation) varies as a function of distance between image pixels. There are, in general, two approaches to extracting texture features using geostatistical methods: 1) using the actual values of the sample variograms directly (Berberoglu et al. Citation2007; Carr and De Miranda Citation1998; Maillard Citation2003); or 2) using coefficients extracted from modelled variograms (Balaguer et al. Citation2010; Herzfeld and Higginson Citation1996). Traditionally, both approaches extract texture features individually from a single band and thus information only from the selected band or bands is used. However, these methods can be extended using multivariate variograms in multispectral dimensions (Li, Cheng, and Guo Citation2009). In a recent study, it was found that cross-variograms derived from different bands of the same image or same bands from different images can also improve classification results when seasonal/temporal information is important (Rodriguez-Galiano et al. Citation2012). In addition, compared to conventional per-pixel methods, sample variograms (and thus the texture features) can also be extracted from image objects/fields (Balaguer et al. Citation2010). This object-oriented approach was found to be more effective for characterizing spatial variations in certain landscapes (e.g. Mediterranean land cover) where between-field spatial variations confound the within-field spatial variations that is of interest (Lloyd et al. Citation2004).

2.1.1.3. Fractal approach

Fractal geometry is used to describe the irregular and highly complex shapes of natural objects such as coastlines and landscapes, which the traditional Euclidean geometry cannot analyse. The fractal dimension (often denoted D), is a non-integer value that measures the irregularity of complex objects. A perfectly flat surface has a fractal dimension of 2.0 and the value of D will increase and approach 3.0 as the surface begins to fill a volume. Therefore, features in a remotely sensed image usually have a fractal dimension of any number between 2.0 and 3.0. Based on the assumption that pixels having the same fractal dimension share the same physical properties, fractal dimensions have been used as a texture measure in remote-sensing image classification (De Jong and Burrough Citation1995; Emerson, Lam, and Quattrochi Citation2005; Sun Citation2006). However, there is a growing consensus that methods used to derive fractal texture images suffer from many limitations and drawbacks, and fractal dimension alone is not sufficient to provide complete texture information (Myint Citation2003; Parrinello and Vaughan Citation2002; Sun et al. Citation2006). One example is that the moving window used for local fractal dimension calculation must be large enough to perform a regression while small window sizes are desired for better classification accuracies. Silvetti and Delrieux (Citation2013) proposed a solution to this problem by computing the local fractal dimension based on quadratic self-correlation and showed that this improved method worked well even with very small windows. Similarly, combining other texture measures or using multifractal dimensions were also proposed to overcome the limitation of using one single fractal dimension. Pant, Singh, and Srivastava (Citation2010) used both fractal dimension and spatial autocorrelation statistics (Moran’s I) and the combined texture image yielded better classification results. Improved classification accuracies were also achieved by adopting multi-fractal textural descriptors, since multi-fractal measures consider the whole spectrum of dimensions and thus are able to provide a higher level of textural information (Parrinello and Vaughan Citation2006; Wawrzaszek, Krupinski, and Aleksandrowicz. Citation2013).

2.1.1.4. Filter-based approach

Although filter-based approaches have been extensively applied for image interpretation in computer science, it was not until recently that these methods were introduced to the field of remote sensing for texture feature extraction purposes (Jain and Healey Citation1998). Gabor filters and wavelet transformation, among many other image-filtering methods, are the two methods most frequently adopted in the remote-sensing literature. Both methods are well known for their ability to characterize image texture at multiple scales by incorporating either different spatial frequencies (Gabor filters) or multiple levels of decomposition (wavelet transformation) (Angelo and Haertel Citation2003; Myint, Lam, and Tyler Citation2004; Zhu and Yang Citation1998). However, textures were extracted only from single-band grey-level images in early applications. The introduction of the opponent features allowed Gabor filters to characterize texture information across multiple spectral bands (Jain and Healey Citation1998; Shi and Healey Citation2003). Similarly, Huang, Zhang, and Li (Citation2008) extracted both spectral and spatial feature sets from remote-sensing images by applying separate wavelet transformation on multispectral bands and PCA band 1. More recently, the spectral dimension was added to traditional spatial dimensions in both filters. It was found that, when image filtering was conducted in the spectral-spatial domain, both the 3-D Gabor filters (Bau, Sarkar, and Healey Citation2010; Shen and Jia Citation2011; Shen et al. Citation2013) and 3-D wavelet transformation (Guo, Huang, and Zhang Citation2014; Huang and Zhang Citation2012) performed better than traditional 2-D approaches, as the new methods were capable of capturing correlations in space, wavelength, and joint spatial/spectral correlation.

2.1.2. Morphological image analysis

Another set of methods that can incorporate spatial information at the pre-classification stage is morphological image analysis, or mathematical morphology (MM) (Soille and Pesaresi Citation2002). MM can extract or suppress desired image objects by using set operators (such as union, intersection, and complementation) with a set of carefully selected (for both shape and size) structuring elements (SEs). The two most widely used morphological operators are opening and closing, which are combinations of the fundamental erosion and dilation operators. Aside from deleting objects smaller than the SEs, opening and closing also modify or deform objects. To avoid this problem and preserve the shape of objects, reconstruction is often used. Opening or closing by reconstruction preserves the shapes of image objects (pixels and their neighbouring pixels) by using connected operators (such as geodesic operators). However, Bellens et al. (Citation2008) showed that due to the fact that many connected simple objects in remote-sensing images (such as roads and buildings) are treated as a single object by the reconstruction process, this often leads to unexpected and undesirable results. This problem of ‘over-reconstruction’ can be solved by partial reconstruction, which limits the extent of reconstruction (Liao et al. Citation2012).

When applying MM in remote-sensing applications, it is also important to include a range of SEs with different sizes to characterize different structures present in the image. This can be addressed by using morphological profiles (MPs), which are defined using successive opening/closing operations with an SE of increasing size (Pesaresi and Benediktsson Citation2001). The resulting MP consists of both an opening profile and a closing profile with a range of different SE sizes, and thus is a powerful tool for extracting spatial information from remotely sensed images. Nevertheless, MP cannot be used in multispectral dimensions. The most commonly adopted method of MP construction for multispectral images is the Extended Morphological Profile (EMP) (Benediktsson, Palmason, and Sveinsson Citation2005). An EMP is constructed by building MPs using spectral features extracted from the original image (via feature extraction methods such as principle component analysis or independent component analysis). Therefore, an EMP contains both spectral information from the extracted features and spatial information enclosed in the MPs. Classification accuracy of high spatial resolution hyperspectral data using EMPs increased significantly compared to traditional spectral classification (Fauvel et al. Citation2008, Citation2013). More recently, in order to overcome limitations such as the fixed shape of SEs, the lack of ability to describe grey-level-related information, and the high computational complexity of the MPs and EMPs (Ghamisi, Benediktsson, Cavallaro, et al. Citation2014), Dalla Mura et al. (Citation2010) proposed the concept of morphological attribute profiles (APs). In contrast to morphological filters used in MPs, APs are constructed by the sequential application of morphological attribute filters (AFs). As a result, image decomposition (and thus attributes of a connected component) can be based on any type of measurement, such as scale (which is used in MPs), shape, and texture (Ghamisi, Benediktsson, and Sveinsson Citation2014; Ghamisi, Mura, and Benediktsson Citation2015). Similar to EMPs, extended APs (EAPs) have been used to handle multispectral/hyperspectral images.

2.1.3. Other spatial information extraction methods at pre-classification stage

Spatial autocorrelation or spatial statistics, such as Moran’s I and Geary’s C, have been used in several studies to characterize spatial correlation among pixels in remote-sensing images (Emerson, Lam, and Quattrochi Citation2005; Myint Citation2003; Myint and Lam Citation2005). These statistics are often calculated based on the spectral variability of the pixels within a predefined window (with a certain size and shape), making it difficult to extract geometric information regarding image objects. Chen, Qin, et al. (Citation2014) proposed a directional spatial correlation (DSC), based on Moran’s I and Geary’s C, to characterize the size, shape, and spectral variability of image objects. The algorithm delineates the geometry of an object by expanding from the centre pixel along eight different directions with various lengths. Therefore, objects of different shapes and sizes can be characterized by the calculated measures. Similarly, Zhang et al. (Citation2006) defined a pixel shape index (PSI) for each image pixel to describe shape features by using extended direction lines around a central pixel. The performance of PSI was later further improved by Huang, Zhang, and Li. (Citation2007) using a structural feature set (SFS) extracted from the directional lines histogram (including PSI).

2.2. Sample selection

It was established many years ago that spatial autocorrelation should be taken into account when sampling reference data, which is then used as either training or validation data (Congalton Citation1991; Campbell Citation1981). Because geographical data have spatial autocorrelation, the assumption of independence of the samples may be violated if groups of contiguous pixels are selected. Therefore, the design of sample unit, sample size, and sampling scheme should account for spatial information.

2.2.1. Sample unit

The potential sampling units have been grouped into three categories: 1) a single pixel; 2) a cluster of pixels; and 3) an object (Congalton and Green Citation2009; Chen and Stow Citation2002). When the area of each sample unit increases, the positional error has diminished impact on the reported accuracy.

Single pixels are used as training samples in order to avoid the effects of spatial autocorrelation. However, it has been found that using single pixels does not guarantee a higher classification accuracy than using clusters of pixels. On the other hand, there may be the disadvantage of under-representing the spatial information for spatially heterogeneous classes (Chen and Stow Citation2002). When used as validation samples, single pixels have also been preferred less because of the difficulty in registering the ground-collected samples with the classification image at single-pixel accuracy (Congalton and Green Citation2009).

The spatial autocorrelation of the training data, using clusters of pixels, was found to decrease when the spatial resolution of the image was coarser (Chen, Stow, and Gong Citation2004). Small clusters of pixels have been found more effective in representing the spatial and spectral information for classes with high spatial heterogeneity (Chen and Stow Citation2002). The validation data can also be collected at a unit of clustered pixels (e.g. 3 pixels × 3 pixels) instead of a single pixel to mitigate the bias in the estimate of the accuracy caused by the spatial misregistration between the classification result and the validation data (Foody Citation2008). The pixels within each cluster or block are all used as training samples. A sensitivity analysis by Mountrakis and Xi (Citation2013) confirmed that these spatially grouped validation samples can achieve similar classification precision to that of the spatially random samples.

It has been claimed that for object-based classification, objects, but not pixels, should be used as sampling units (Congalton and Green Citation2009; Stehman and Wickham Citation2011). Nevertheless, Radoux et al. (Citation2011) after reviewing 20 OBIA-related papers found that only 40% of them used objects as sampling units, and that a majority of research are still using pixels as sampling units.

Objects as sample units have advantages for object-based classification. First, the sample unit corresponds to the mapping units. Second, potential error due to spatial misregistration is greatly reduced (Congalton and Green Citation2009). On the other hand, when objects are used as sample units for validation data, there are some critical issues: 1) heterogeneity may exist within a sample unit (Stehman and Wickham Citation2011); and 2) the sample units in one validation data set may have various sizes and thus areas (Radoux and Bogaert Citation2014). Accuracy assessment methods aiming to resolve these issues have been developed (Radoux et al. Citation2011; Lizarazo Citation2014). It has been found that larger objects are better classified (Radoux and Bogaert Citation2014; Castilla et al. Citation2014). Therefore, it is important to have objects of different sizes in the validation sample set (1029–1037).

2.2.2. Sample size

The optimal sample size of the training data is the minimum sample size that is required to achieve a certain classification accuracy (Foody et al. Citation2006). (Chen and Stow Citation2002), after conducting experiments using images of different spatial resolution levels (2 m, 4 m, 8 m, 12 m, and 16 m), found a positive relation between the required number of training samples and spatial heterogeneity.

The optimal sample size of the validation data is the minimum sample size that is required to estimate the classification accuracy with a certain precision (Congalton and Green Citation2009; Baraldi, Boschetti, and Humber Citation2014). Traditional algorithms for theoretically determining the optimal sample size are mainly based on binomial or multinomial distributions, aiming at reaching a pre-specified class-specific accuracy and confidence interval (Congalton Citation1991; Baraldi, Boschetti, and Humber Citation2014). Nevertheless, the optimal sample size is dependent on the classification scenario and the sampling design. It has been found that the optimal sample size for object-based accuracy assessment is smaller than that of pixel-based accuracy assessment (Radoux et al. Citation2011; MacLean et al. Citation2013).

2.2.3. Sampling scheme

Although probability sampling protocols such as simple random sampling or systematic random sampling are readily available, in practice it is common to select samples from specific areas where the true label of the ground is known (Zhen et al. Citation2013).

Stehman (Citation2009) suggested seven criteria for the sampling sheme for selecting reference samples. One of these is that the samples should be evenly distributed in space in order to represent the entire study area. To ensure spatial balance, a common protocol for sampling, especially when the study area is large, is to first define spatially evenly distributed subareas, and then randomly select sample units from each of the subareas (Zhao et al. Citation2014; Dihkan et al. Citation2013).

A typical approach is to separate a reference data set into training and validation samples (Mountrakis and Xi Citation2013). In that case, it’s important to consider the spatial dependence between the training and validation samples. If both data sets are randomly selected within the same patches, there tends to be an optimistic bias (Zhen et al. Citation2013). Therefore, it is recommended to separate the known patches into training patches and validation patches before random sampling.

2.3. Classifiers

2.3.1. Object-based image analysis

Object-based image analysis (OBIA) paves the way for combining spectral and spatial information, and in doing so potentially offers a more comprehensive classification approach. It seeks to create meaningful image objects by segmenting an image into groups of adjacent pixels with similar characteristics (e.g. spatial and spectral). It uses image objects rather than just individual pixels for image analysis (Benz et al. Citation2004; Hay and Castilla Citation2008). With the advent and proliferation of the availability of high-resolution imagery, the use of OBIA has increased rapidly (Blaschke Citation2010; Blaschke et al. Citation2014; Liu, Guo, and Kelly Citation2008). OBIA primarily encompasses two stages: image segmentation and classification. In the image segmentation process, an image is partitioned into non-overlapping homogeneous regions (i.e. image objects) based on one or more criteria of homogeneity (e.g. colour and shape in the software package eCognition) (Fu and Mui Citation1981; Tarabalka et al. Citation2012; Blaschke Citation2010). In general, within-object variance is minimized compared to between-object variance (Desclée, Bogaert, and Defourny Citation2006). Each segmented image object can be characterized by both spectral and spatial properties (Jensen Citation2000; Benz et al. Citation2004; Blaschke and Strobl Citation2001; Baatz and Arno Citation2000). Spectral properties of an image object are typically statistical measures of spectral values, such as the pixels’ mean, standard deviation, and spectral ratios. Spatial properties include the shape (e.g. area, length, compactness, and smoothness), neighbourhood, and topologies. Spatial properties explicitly embodied by the individual image object are crucial to OBIA and are the main reason for the marked increase in OBIA-related studies in recent years (Hay et al. Citation2003; Blaschke et al. Citation2014; Blaschke, Lang, and Hay Citation2008; Hay and Castilla Citation2006; Conchedda, Durieux, and Mayaux Citation2008). During the classification stage, homogeneous image objects, instead of individual pixels, are analysed based on their spectral and spatial properties using traditional parametric or machine learning classification approaches, knowledge-based approaches, or fuzzy logic approaches (Jensen Citation1996; Jensen Citation2000; Benz et al. Citation2004). An iterative application of segmentation and classification is usually desirable in OBIA for extracting the information of interest (Benz et al. Citation2004). The initial classification of segmented image objects can serve as the high-level input for subsequent segmentation (i.e. classification-based segmentation) to improve the classification accuracy.

Over the last decade, a variety of segmentation methods have been developed to partition an image into relatively homogenous image objects (Fauvel et al. Citation2013; Pal and Pal Citation1993; Neubert, Herold, and Meinel Citation2008). From an algorithmic perspective, these methods can be categorized into three groups: point-based, edge-based, and area-based (Pal and Pal Citation1993; Schiewe Citation2002). Depending on the features employed in the segmentation process, the methods can be categorized into the following groups: spatial-based, spectral-based, and spatial-spectral-based (Fauvel et al. Citation2013; Fu and Mui Citation1981). Spatial-based methods search for homogenous regions of spatially connected pixels with the defined criteria (e.g. greyscale texture). Representative methods in this group include region growing, split-merge methods, and watershed methods (Blaschke, Burnett, and Pekkarinen Citation2004; Gonzalez, Woods, and Eddins Citation2004). Spectral-based methods, such as thresholding and partitional clustering, group pixels into spectral clusters based on spectral similarity measures, without considering the spatial locations of these pixels. The disjointed regions within the same cluster are assigned different labels (Fu and Mui Citation1981; Fauvel et al. Citation2013; Tarabalka, Benediktsson, and Chanussot Citation2009; Celeux and Govaert Citation1992). Spatial-spectral-based methods exploit both spatial and spectral information in the segmentation process, and allow for the merging of spatially disjoint regions (Tilton Citation1998). One example is HSeg segmentation, which is based on hierarchical stepwise optimization (HSWO) (Beaulieu and Goldberg Citation1989; Tilton et al. Citation2012). HSeg segmentation combines region growing with unsupervised classification to merge the pre-segmented image objects in terms of spectral similarity measures iteratively, and generates a hierarchical sequence of image objects with different levels of detail (Fauvel et al. Citation2013; Plaza et al. Citation2009; Tarabalka et al. Citation2012; Tilton et al. Citation2012). Technical details of image segmentation methods are reviewed in Pal and Pal (Citation1993), Fu and Mui (Citation1981), and Ilea and Whelan (Citation2011) .

One critical parameter in implementing the segmentation is scale (Blaschke Citation2010; Benz et al. Citation2004; Burnett and Blaschke Citation2003; Wang, Sousa, and Gong Citation2004). The larger the scale value, the larger the image objects grow in the segmentation (Baatz and Schäpe Citation1999). Though a single scale (i.e. one-level representation) might be sufficient or straightforward for representing the content in specific applications (Tiede, Lang, and Hoffmann Citation2008), it is often not possible to determine a single appropriate scale in advance. Besides, image objects of interest usually have their own inherent scales. Significant image objects may be generated at different scales of analysis (Hay et al. Citation2001, Citation2003; Woodcock and Strahler Citation1987). Multi-scale representation is thus more appropriate for scene understanding and image interpretation. However, this multi-scale representation cannot be achieved simply by changing the resolution of the imagery, since this results in the loss of much useful context information (Benz et al. Citation2004). OBIA conducted at multiple scales can overcome this limitation to a large degree. Hence, establishing a hierarchical multi-scale image object network is important for appropriately interpreting the phenomena and understanding relations within an image (Hay and Castilla Citation2008; Burnett and Blaschke Citation2003). This hierarchical image object network can be built with eCognition, which builds on fractal net evolution segmentation and is currently the most widely used object-based image analysis system (Benz et al. Citation2004). Through defining the scale range and homogeneity criteria, image objects can be created at multiple scales in eCognition wherein each successive scale contains larger segmented image objects (Baatz and Arno Citation2000; Baatz et al. Citation2000; Kim et al. Citation2011). To date, a significant number of research studies (e.g. 50–55% of more than 800 articles identified by Blaschke (Citation2010)) were conducted with eCognition software. Yet the strong dependence of the OBIA research on eCognition software dilutes the influence of various alternative hierarchical segmentation approaches on object classification accuracy. A wide spectrum of hierarchical segmentation approaches (e.g. hierarchical split merge refinement (HSMR), multi-scale object-specific analysis (MOSA), and HSeg) need to be investigated and evaluated to construct a more comprehensive object-based image analysis system (Hay et al. Citation2001; Tilton et al. Citation2012; Ojala and Pietikäinen. Citation1999).

At each segmentation scale, the image object in the network is characterized by its spectral-spatial features and neighbours. It is also topologically linked to its sub-objects and super-objects in a strict hierarchical structure (Benz et al. Citation2004; Mallinis et al. Citation2008). The hierarchical scale dependencies among the image objects allow for understanding scaling relations and exploring substructures of general land-cover regions. Methods for upscaling image objects hierarchically have been investigated, specifically for evaluating whether segmented image objects created at various scales should be built precisely one upon another, or generated independently with the delineated outlines fitting exactly (Blaschke Citation2010; Hay et al. Citation2005, Citation2001; Castilla, Hay, and Ruiz-Gallardo Citation2008). Owing to the hierarchical relations, different procedures can be implemented in different types of super-objects (e.g. urban and agricultural areas) to more appropriately characterize the landscape structure and explain the phenomenon (Benz et al. Citation2004). Moreover, the hierarchical relationship between sub-objects and super-objects can be used to determine the validity of sub-object classes (e.g. pasture), given the super-object type (e.g. urban areas). As the magnitude of aggregation and abstraction, scale is a crucial aspect of image understanding and a hot research topic in OBIA (Blaschke Citation2010; Blaschke et al. Citation2014).

Geographic object-based image analysis (GEOBIA) has recently been perceived as a new and evolving paradigm in remote sensing and geographic information science (GIScience) (Blaschke et al. Citation2014). Compared to OBIA, GEOBIA emphasizes the geographical space of image objects and the integration of image objects with GIS applications. GEOBIA provides a spatial perspective for understanding various physical and abstract image objects (e.g. different agricultural systems in different locations), with the ultimate goal of developing spatial knowledge or geographic-based intelligence (Blaschke et al. Citation2014; Blaschke Citation2010). GEOBIA typically involves the human semantics, ontologies, and perceptions as guiding principles for describing complex scene content, and conceptualizing knowledge based on meaningful image objects within a geographical context (Lang Citation2008). This spatial expert knowledge incorporation seems to be an increasing trend in remote sensing and GIScience (Blaschke et al. Citation2014; Andres, Arvor, and Pierkot Citation2012; Yue et al. Citation2013), but might not be so obvious to computer vision, signal processing, and pattern recognition research that focuses on advancing segmentation or classification algorithms. More cooperation across disciplines needs to be encouraged to promote the development of this multidisciplinary research field.

2.3.2. Pixel-based classifiers

In recent years, spatial information has been increasingly explored in the pixel-based classification process, especially for high-resolution imagery. Most methods investigate the role of spatial information (e.g. texture) as additional feature bands without modifying the classifiers. Contextual classifiers, however, exploit the spatial information among neighbouring pixels in classifiers to improve the classification results. Two representative methods, namely contextual support vector machines and Markov random fields, are discussed below. Sparse representation, which has recently been explored as a method to incorporate spatial information, is then reviewed. Finally, as a means to quantify the spatial structure and dependence, geostatistical methods are also reviewed.

2.3.2.1. Contextual support vector machines

Over the last decade, intensive efforts have been conducted to advance the performance of classifiers in remote sensing (Camps-Valls et al. Citation2014; Fauvel et al. Citation2013; Plaza et al. Citation2009). Among all classifiers, support vector machines (SVMs) are particularly appealing to the remote-sensing community (Mountrakis, Im, and Ogole Citation2011). SVMs have generally been found not to be susceptible to the Hughes phenomenon (i.e. a degradation of classification performance with increasing numbers of input variables) and have shown remarkable performance in numerous applications, especially with limited training samples. A thorough review can be found in Mountrakis, Im, and Ogole (Citation2011). As a non-parametric pixel-wise statistical learning technique, SVM aims to construct an optimal separating hyperplane that maximizes the margin (i.e. space) between classes based on support vectors. Support vectors are defined as the nearest vectors from each class to the hyperplane (Suykens and Vandewalle Citation1999; Vapnik Citation2013; Cortes and Vapnik Citation1995). When data cannot be linearly separated, SVM solves the inseparability problem with kernel functions, which project the data into a higher dimensional feature space (Bioucas-Dias et al. Citation2013; Cristianini and John Citation2000; Schölkopf and Smola Citation2002). The commonly used kernel functions in SVMs are Gaussian radial basis function (RBF), polynomial, linear, and sigmoid functions (Chang and Lin Citation2011; Schölkopf and Smola Citation2002). Conventionally, these kernels are employed to project the spectral vectors of pixels into the higher dimensional feature space, without explicitly considering the spatio-contextual information.

Owing to the properties of kernel functions (e.g. a scaled summation of Mercer’s kernels are valid kernels), several SVM variants have recently been developed to simultaneously exploit the spatial and spectral information (Plaza et al. Citation2009; Camps-Valls et al. Citation2014; Fauvel et al. Citation2013). One approach is composite kernel-based SVM, which includes a weighted linear combination of spectral and spatial kernels computed with available features (Camps-Valls et al. Citation2006; Fauvel, Chanussot, and Benediktsson Citation2012; Fauvel et al. Citation2013). Spatial kernels define the projection rule of pixel-based spatial features, which can be derived with statistical filtering (e.g. mean, median, and variance) applied to each pixel’s neighbourhood (e.g. morphological neighbourhood or fixed neighbourhood) (Fauvel et al. Citation2013; Fauvel, Chanussot, and Benediktsson Citation2012). Compared to stacked spatial and spectral information as feature bands in SVM, the composite kernel allows specification of the relative influence of the extracted features, and offers more flexibility for integrating diverse information in land-cover mapping (Fauvel, Chanussot, and Benediktsson Citation2012; Camps-Valls et al. Citation2014). However, spatial kernels in the composite currently only summarize basic information (e.g. median) of the spatial neighbourhood. More sophisticated techniques for describing spatial structures (e.g. shape, size, and homogeneity) of the neighbourhood are called for (Fauvel, Chanussot, and Benediktsson Citation2012). Recently, this composite kernel has been further extended to multiple kernel learning, which provides a kernel framework for fusing multi-source spectral and spatial information in SVM classification (Tuia et al. Citation2010a, Citation2010b).

The introduction of composite kernels opens a wide range of opportunities for exploring the use of spatial information in classifiers. Instead of extracting spatial features before kernel computation, a recent advance in SVM is the use of graph kernels to capture all possible spatial relations between pixels of a neighbourhood in the graph-based feature space (Camps-Valls, Shervashidze, and Borgwardt Citation2010). Through working directly with the original data, graph kernel-based SVM overcomes the limitations associated with explicit spatial feature computation and is also advantageous in exploiting spatial relations at multiple scales.

2.3.2.2. Markov random fields

Markov random fields (MRFs) are a powerful family of stochastic models for exploring the spatial dependence properties associated with images on a per-pixel basis. They generalize the concepts of a 1-D Markov chain to a 2-D framework (e.g. pixel lattices) for characterizing contextual information and exploiting the continuity among neighbouring pixels in a classifier (Moser, Serpico, and Benediktsson Citation2013; Plaza et al. Citation2009). In the Bayesian decision theory framework, the class label of each pixel is predicted through the Bayesian maximum a posteriori (MAP) rule, which optimizes the parameters in the framework by maximizing the posterior mass functions (Moser, Serpico, and Benediktsson Citation2013; Li Citation2009; Kindermann and Snell Citation1980). However, due to the large number of parameters to be optimized jointly, this global MAP image classification is computationally intractable.

MRFs offer a computationally feasible solution to this issue by assuming that all the pixels satisfy the Markovianity condition. This condition implies that the probability distribution of the class label of each pixel is only affected by the labels of its neighbouring pixels (Li Citation2009; Geman and Geman Citation1984). With this assumption, MRFs can significantly reduce the computational burden in optimizing the parameters in the Bayesian framework, by changing from a global model complying with the MAP rule, to a local model based on its neighbourhood system. MRF-based approaches have therefore gained increasing popularity in remote sensing over the last decade (Li Citation2009; Moser, Serpico, and Benediktsson Citation2013; Melgani and Serpico Citation2003; Tso and Olsen Citation2005; Jia and Richards Citation2008).

In MRFs, based on the Hammersley–Clifford theorem, the Bayesian MAP rule can be formulated as the minimization of the energy function (Geman and Geman Citation1984; Dubes and Jain Citation1989). The basic energy function can be expressed as a linear combination of two parts. One part is related to the conditional pixel-wise spectral information, and it is the same as the non-contextual Bayesian classification system. The second part (also known as Potts model or multi-level logistic model) considers the spatio-contextual information of each pixel with the Kronecker function, which favours the generation of connected regions in spatially homogeneous land-cover classes (Jackson and Landgrebe Citation2002; Moser, Serpico, and Benediktsson Citation2013; Geman and Geman Citation1984). To minimize the energy function in random fields, two classical iterative approaches have been adopted in the literature, namely simulated annealing (SA) and iterated conditional mode (ICM) (Geman and Geman Citation1984; Solberg, Taxt, and Jain Citation1996; Moser, Serpico, and Benediktsson Citation2013). However, the long computation time of SA and the local convergence issue associated with ICM limit their abilities in energy minimization. A number of global and local minimization methods, including message passing (e.g. loopy belief propagation and tree-reweighted message passing), convex relaxation, and semi-global labelling, have been proposed for tackling the optimization problem (Szeliski et al. Citation2008; Yedidia, Freeman, and Weiss Citation2003; Kolmogorov Citation2006; Wainwright, Jaakkola, and Willsky Citation2005; Schindler Citation2012). One approach that has recently drawn considerable attention is graph-cut (Kolmogorov and Zabih Citation2004; Boykov and Funka-Lea Citation2006; Greig, Porteous, and Seheult Citation1989; Boykov, Veksler, and Zabih Citation2001). This method minimizes the energy function through solving a maximum flow problem over a graph. Graph-cut methods have been demonstrated to converge to the global minimum for two classes, or a strong local minimum for more than two classes, with acceptable computation time (Kolmogorov and Zabih Citation2004).

MRFs exhibit considerable capability in incorporating multiple spatio-contextual information sources in the classification process. The information is not limited to the spatial dependence among neighbouring pixels, but also involves spatial edges, geometrical structures, multi-scale analysis, and spatial anisotropy (Moser, Serpico, and Benediktsson Citation2013; Li Citation2009; Solberg, Taxt, and Jain Citation1996; Moser and Serpico Citation2008). The energy function in MRFs can be extended to a linear combination of multiple contribution parts, with each information source being one part. This flexibility in the choice of energy function makes MRFs a powerful tool for meeting the specific requirements of complex land-cover classification systems, and taking into account the various roles of spatial information in diverse applications to improve classification accuracy (Moser, Serpico, and Benediktsson Citation2013; Moser and Serpico Citation2008, Citation2011). MRFs have therefore demonstrated advantages over other pixel-based classification methods. A current challenge with MRFs is to develop more novel and reliable methods, such as the integration of MRFs with SVMs, to extract accurate remotely sensed information (Zhang et al. Citation2011; Moser and Serpico Citation2013). Investigating new input information sources (e.g. new types of spatial information) also plays a crucial role in advancing this spatio-contextual classifier in land cover mapping.

As a generative model, MRFs comply with the MAP rule through modelling the joint probability of the observed data and the corresponding class labels (i.e. the product of class conditional probability (also named likelihood) and class prior probability) (Li Citation2009). MRFs need to explicitly model the distribution of the likelihood and assume that the observed image data are conditionally independent given the class labels (Kumar and Hebert Citation2003; Besag Citation1974). This conditional independence assumption makes the computation tractable, but meanwhile limits the contextual information to the labelled image (Kumar and Hebert Citation2003; Zhong and Wang Citation2010). To directly take into account the spatio-contextual information in both observed and labelled images, conditional random fields (CRFs) have been proposed and recently utilized in image classification (Lafferty, McCallum, and Pereira Citation2001; Wallach Citation2004; Kumar and Hebert Citation2003; Zhong and Wang Citation2010). CRF assumes that class labels, conditional on the observed data, obey the Markov property. As a discriminative model, CRF directly models the posterior probability of class labels, given the observed image data, as the Gibbs distribution. It avoids the problem of explicit modelling of the likelihood in MRFs and relaxes the corresponding conditional independence assumption. CRFs can be used to classify various remotely sensed images regardless of the distributions they follow, and are capable of incorporating the spatial contextual information in both observed and labelled images (Zhao, Zhong, and Zhang Citation2015; Zhong, Zhao, and Zhang Citation2014; Kosov, Rottensteiner, and Heipke Citation2013; Hoberg et al. Citation2015; Zhong, Lin, and Zhang Citation2014).

Similar to MRFs, the maximization of the posterior probability in CRFs can be formulated as the minimization of the energy function. The most commonly used CRF energy function in image classification can be written as the sum of the unary and pairwise potentials (Kumar and Hebert Citation2003). The unary potential is formulated to model the pixel-wise relationship between the spectral information of each pixel and its corresponding class label without considering its neighbours. The minimization of the energy that consists of only this potential is equivalent to the non-contextual Bayesian classification of the image. Generalized linear models (e.g. multinomial logistic regression) have been utilized in the unary potential to model the class posteriors given the observations (Kumar and Hebert Citation2003; Zhong and Wang Citation2010). Yet the unary potential is not restricted to a specific form. Recently, state-of-the-art discriminative classifiers, such as support vector machines, Gaussian mixture models, and random forests, have also been explored (Kosov, Rottensteiner, and Heipke Citation2013; Schindler Citation2012; Zhong, Lin, and Zhang Citation2014; Zhao, Zhong, and Zhang Citation2015). The pairwise potential is used to model the spatial contextual information of each pixel in terms of its neighbourhood in both observed and labelled images. The pairwise potential learns a data-dependent pairwise smoothness term (e.g. generalized Ising model) with an assumption of smoothness to favour the neighbourhood pixels with similar class labels in homogeneous image regions (Zhong and Wang Citation2010; Kumar and Hebert Citation2003). It takes into account the neighbouring class labels and the whole observed image data in determining the class label of each pixel. CRFs have been demonstrated to perform better than non-contextual classification methods (e.g. random forests), but may oversmooth the image with the relatively simple smoothness term. More complex and sophisticated smoothness functions thus need to be investigated in CRFs (Schindler Citation2012). Recently, distance boundary constraints have been explored in the pairwise potential to preserve the spatial details in the image classification result (Zhang and Jia Citation2012; Zhong, Lin, and Zhang Citation2014). Additionally, CRFs can incorporate high-order potentials, which can model more wide-ranging and complex contextual information with the high-order spatial neighbourhood (Zhong and Wang Citation2011). Yet these high-order potentials increase the complexity of models and make efficient inference problematic.

2.3.2.3. Sparse representation

Sparse representation has recently been investigated in classifying remotely sensed imagery. In sparse representation, each pixel in the image is assumed to be represented by a sparse linear combination of training samples from a dictionary (Chen, Nasrabadi, and Tran Citation2011; Zhang et al. Citation2014b; Song et al. Citation2014). The dictionary can be designed based on the mathematical model of the data, or a set of realizations of the data (Rubinstein, Bruckstein, and Elad Citation2010). The coefficients associated with training samples in sparse representation are the sparse vector, whose non-zero entries denote the weights of the selected training samples. The sparse vector can be recovered through sparsity-constrained optimization, which aims to minimize the approximation error between the observed and reconstructed pixel values, given the sparsity level constraint. The constraint of the sparsity level is used to set the upper bound of the number of non-zero entries in the sparse vector. In sparse representation, the spatial contextual information from the neighbouring pixels can be exploited through imposing another constraint (i.e. smoothness constraint) in the optimization process. For example, two types of smoothness constraints, namely the Laplacian constraint and the joint sparsity model, have been devised within the sparsity-constrained optimization formulation (Chen, Nasrabadi, and Tran Citation2011). The Laplacian constraint forces the reconstructed pixel of interest to share similar spectral characteristics with those of its four nearest neighbours, while the joint sparsity model assumes that the pixels in a small neighbourhood can be modelled simultaneously by the sparse linear combination of a few common training samples. The sparse vector recovered through optimization can then be utilized to determine class labels of pixels of interest. Recently, kernel sparse representation has been proposed to represent pixels in terms of sparse linear combination of training samples in a feature space induced by a kernel function (Chen, Nasrabadi, and Tran Citation2013; Liu et al. Citation2013). In the optimization formulation, the kernelized joint sparsity model, for example, is served as the smoothness constraint in the induced feature space to incorporate the contextual information across neighbouring pixels (Chen, Nasrabadi, and Tran Citation2013). Kernel sparse representation has been demonstrated to improve data separability and outperform the conventional linear sparse representation.

2.3.2.4. Geostatistical methods

Rooted in variography, geostatistical techniques offer a valuable means to quantify the spatial image structure and spatial dependence between observations (Curran and Atkinson Citation1998; Atkinson and Lewis Citation2000). Variograms are key functions in geostatistics as they are essential in characterizing the spatial variance structure, with the stationary assumption, and play a crucial part in other geostatistical techniques (e.g. kriging and stochastic simulation) (Atkinson and Lewis Citation2000; Van Der Meer Citation2012). In recent years, variogram-based texture measures from images have been increasingly used as additional feature bands to be incorporated into the classification process (Atkinson and Lewis Citation2000; Van Der Meer Citation2012). Geostatistical techniques, however, are still rarely used as classifiers or spatial variants in classifiers (Atkinson and Naser Citation2010). Among the few studies conducted, indicator kriging has been used to exploit the spatial variability of classes within the image and estimate class probabilities for each pixel (Van Der Meer Citation1994). This geostatistical method has been shown to be more accurate than conventional classification approaches (Das and Singh Citation2009; Park, Chi, and Kwon Citation2003). Additionally, geostatistically weighted K-nearest neighbour (k-NN) has been employed to incorporate the spatial variation between classes for classifying remotely sensed imagery. This classifier modifies the original k-NN with three geographical weightings and allows for the integration of the feature information from both spatial and spectral domains (Atkinson and Naser Citation2010; Atkinson Citation2004).

2.4. Post-classification

Spectral-based pixel-wise classified images usually suffer from salt-and-pepper noise, due to the lack of spatial information in the classification (Lu and Weng Citation2007). Spatio-contextual information can be incorporated into the post-classification process to reduce the noise effect and build more spatially consistent thematic maps (Fauvel et al. Citation2013). The spatio-contextual post-processing typically consists of two elements: spatial neighbourhoods and spatial regulation. Spatial neighbourhoods can be obtained through image segmentation, or simply arbitrarily defined (e.g. first- and second-order neighbourhoods) (Tarabalka, Benediktsson, and Chanussot Citation2009; Fauvel et al. Citation2013; Tarabalka et al. Citation2010b). Spatial regulation can then be performed to modify the label of each individual pixel based on its neighbourhood. The regulation process can help filter out spatially isolated pixels, enhance spatial patterns of classified images, and create more spatially homogenous maps. The most commonly used spatial regulation method is majority voting, which assigns the pixels to the most frequent class in its spatial neighbourhood. Majority voting has been used in several studies to refine pixel-wise classification (e.g. SVMs) results in segmented spatial neighbourhoods (Tarabalka, Chanussot, and Benediktsson Citation2010; Tarabalka, Benediktsson, and Chanussot Citation2009; Tarabalka et al. Citation2010a; Fauvel et al. Citation2013). Additionally, Gaussian smoothing, bilateral filtering, edge-aware filtering, and anisotropic diffusion have been explored as spatial regulation methods to locally smooth classification results, through taking into account the labels, probabilities, and intensity values of the pixels within the spatial neighbourhood (Schindler Citation2012; Huang et al. Citation2014b).

Spatial reclassification kernel (SPARK) has also been employed to regularize classified images (Barnsley and Barr Citation1996). SPARK refines the class of each centred pixel in defined kernels (or fixed window sizes) based on the frequency and spatial arrangements of adjacent land-cover classes. The kernel size for each individual classified pixel in SPARK is further optimized with a similarity index to take into account the spatial variation of different land-cover types (Van Der Kwast et al. Citation2011). SPARK exhibits potentials for distinguishing complex spatial assemblages and subtle differences of land-use types in urban areas (Barnsley and Barr Citation1996). Moreover, relearning methods (i.e. relearning-Hist and relearning-PCM) have recently been proposed to adaptively learn the frequency and spatial arrangement of class labels in the neighbourhood (Huang et al. Citation2014b). Relearning methods have been found efficient in correcting classification errors and enhancing class separability.

MRF-based spatial regulation methods have recently been explored to incorporate spatio-contextual information in the post-classification process. Through minimizing the global energy function, MRF-based spatial regulation methods can iteratively refine the initial spectral pixel-wise classification results to obtain the MAP estimate (Tarabalka et al. Citation2010b). Unlike the original spectral image used in MRF-based classifiers (see Section 2.3.2), the observed data in MRF-based spatial regulation represent the probabilistic pixel-wise classification result (Tarabalka et al. Citation2010b; Huang et al. Citation2014b). The MRF-based spatial regulation method is advantageous, with its flexibility to exploit multiple spatio-contextual information sources (e.g. spatial edge and geometrical structures) (Moser, Serpico, and Benediktsson Citation2013). Tarabalka et al. (Citation2010b) integrated the ‘fuzzy no-edge/edge’ function into the energy function to regularize class memberships while preserving the spatial edge information, and demonstrated that MRF-based spatial regulation methods yield a more accurate classification result compared to other classification methods (e.g. SVM, extraction and classification of homogeneous objects, and watershed segmentation).

2.5. Accuracy assessment

Validation, or accuracy assessment of the classification result, is an essential part of land-cover mapping(Congalton et al. Citation2014).

The accuracy of the classified image has two aspects: geometric accuracy, which is the accuracy related to the geolocation of each pixel or object, and thematic accuracy, which is the accuracy related to the labelling of each pixel or object (Möller et al. Citation2013; Zhan et al. Citation2005). Of more interest to users is thematic accuracy, whereas geometric accuracy has a major impact on the representation of thematic accuracy (Foody Citation2002; Pontius and Millones Citation2011).

2.5.1. Thematic accuracy

2.5.1.1. Object-based assessment

For pixel-based accuracy assessment, one widely recognized routine is to derive the accuracy metrics from the confusion matrix (CM) (Congalton and Green Citation2009; Foody Citation2002). In the case of object-based assessment, the CM can be modified to per-object CM or area-based CM (Lizarazo Citation2014; Stehman and Wickham Citation2011).

The most widely used accuracy assessment metric for object-based image classification is overall accuracy, that is, the ratio of the number of correctly classified sampling units to the total number of selected sampling units, which is the per-object accuracy (Radoux et al. Citation2011; Zhan et al. Citation2005). However, objects are of varying size and area. For large homogeneous areas (e.g. water), to consider the large object as one unit could result in misleading accuracy estimation. Hence, when the sampling unit is an object, it is common to modify the accuracy metric by using the area of the object, instead of the number of objects, in the calculation, which is the area-weighted or area-based accuracy (Stehman and Czaplewski Citation1998; Desclée, Bogaert, and Defourny Citation2006). Another consideration of the object-based accuracy measure is that the object may contain heterogeneity and thus should not be binarized as the ‘correct’ or ‘incorrect’ object (Stehman and Wickham Citation2011). The theme similarity measure proposed by Lizarazo (Citation2014) considers the ratio of overlapping area that has thematic category matching the area of the validation object.

It should be noted that the accuracy derived from validation samples is not a direct estimate of the accuracy of the entire study area. Therefore, based on the idea of area-weighted accuracy, Radoux et al. (Citation2011) and Radoux and Bogaert (Citation2014) proposed modified predictors of accuracy metrics by considering both the sampled and unsampled objects. The correctness of the unsampled objects was estimated using the statistical expectation. They demonstrated that the proposed predictors are also unbiased estimates of the accuracy metrics, but with improved precision.

2.5.1.2. Multi-scale assessment

There have been efforts to assess the accuracy of image classification results at multiple scales. For example, Wardlow and Callahan (Citation2014) assessed the MODIS irrigated agriculture data set at both state level and field level. State-level assessment provides a general summary of the quality of the classification, while per-field level assessment reveals how the classification error is related to crop type and location. Furthermore, they found that fields of larger size were classified with higher accuracy.

Ge et al. (Citation2009) adopted the rough set theory, which is a useful tool for obtaining knowledge from incomplete or inconsistent data, to assess the thematic accuracy of a classification result in a multi-level fashion without the use of validation data. The three scales considered were pixel-level, object/category-level, and map-level. At the pixel-level, Shannon entropy was used to quantify the amount of information required to label a pixel as a particular class with complete certainty. Rough degree and rough entropy, which measure the uncertainty of labelling the object as a certain class, are used to measure object/category-level accuracy. At the image scale or map level, the accuracy and quality of approximation are used to represent the overall uncertainty. An experiment with five different data sets revealed that object/category-level measures have significantly negative correlation with user’s accuracy and class-specific kappa coefficient, indicating the power of the proposed metrics.

2.5.1.3. Geostatistical error assessment

Thematic error is in many cases not randomly distributed in space, but correlated with the position of boundaries between different classes (Foody Citation2002). Geostatistical methods are employed to model the spatial pattern of errors. Kyriakidis and Dungan (Citation2001) proposed a method to quantify local classification accuracy by indicator kriging. Another commonly adopted approach is sequential indicator simulation, which investigates the uncertainty propagation in the classification procedure (Magnussen and de Bruin Citation2003; de Bruin Citation2000). Carneiro and Pereira (Citation2014) extended geostatistical stochastic simulation to integrate the influence of patch size by considering patch size as one source of uncertainty. On the other hand, geostatistical classification frames are also able to provide uncertainty maps as a by-product (Li and Zhang Citation2011b).

2.5.2. Geometric accuracy

Geometric accuracy concerns the accuracy associated with space; the labelling of pixels or objects is of no interest at all. For pixel-based classification, geometric accuracy is commonly referred to as positional accuracy (Congalton and Green Citation2009). For object-based classification, geometric metrics characterizing more properties have been developed, including shape accuracy, edge accuracy, and segmentation accuracy, as well as positional accuracy.

2.5.2.1. Positional accuracy

Pixel-based assessment: due to registration discrepancy, the pixels in the classified image may not coincide perfectly with the validation image or samples. The two-dimensional horizontal accuracy of the classification result can be quantified using the root mean square error (RMSE) between corresponding pixels, which is an overall positional accuracy across the entire image (Congalton and Green Citation2009). The correspondence of the pixels should be determined with interpreted points from both the classified image and the validation image or samples.

Object-based assessment: area-based metrics quantify how much and how well the classified and validation objects overlap. For example, Persello and Bruzzone (Citation2010) considered the situation where more than one object from the classification map overlaps with one validation object. The classified object that has the largest overlapping area with the validation object is considered to be the corresponding one. Another common approach is location-based, quantifying the deviations of the corresponding object centres. For example, Zhan et al. (Citation2005) proposed a quality measure for the position, Q_Loc, which is the Euclidean distance between the centres of mass of the two corresponding objects. Based on Q_Loc, position similarity indices are designed by normalizing Q_Loc and subtracting this from 1. The normalization term used includes the maximum distance from the furthest validation centroid to the centroid of the overlapping area (Möller et al. Citation2013), the square root of the overlapping area (8685–8698), the diameter of a combined area circle (CAC, i.e. a circle that has the same area as the sum of the areas of the two corresponding objects) (Lizarazo Citation2014), and other measures. The RMSE metric can also be applied to object centroids as a global metric to measure the overall positional error of the objects (Whiteside, Maier, and Boggs Citation2014). The other global metrics include the mean and variance of the accuracy metrics for each object (Zhan et al. Citation2005).

2.5.2.2. Shape accuracy

Shape accuracy quantifies the similarity between the shape of a validation object and that of the corresponding classified object. The shape of an object can be described using various well established indices, e.g. compactness, eccentricity, or fractal dimension (McGarigal, Cushman, and Ene Citation2012; Angel, Parent, and Civco Citation2010). Shape accuracy can be estimated from these shape indices. The dissimilarity of the shape indices for the two corresponding objects is the shape-related error, which is also a measure of the accuracy. Ways of comparing the two shape indices include the absolute value difference (Persello and Bruzzone Citation2010), the ratio of the values (Lizarazo Citation2014), and other approaches.

2.5.2.3. Edge accuracy

Edge accuracy quantifies the coincidence between the validation object and the classified object. Usually the ratio of the length of spatially overlapping edges to the length of the validation object edge is calculated to represent the edge accuracy (Persello and Bruzzone Citation2010; Lizarazo Citation2014).

Object fate analysis (OFA) was presented by Schöpfer and Lang (Citation2006) and has been applied to evaluate the deviation of object boundaries (Albrecht, Lang, and Hölbling Citation2010; Hernando et al. Citation2012). For each validation object, six scenarios are defined according to its spatial relationship with the classified objects. Two metrics, extrution and intrusion, are calculated for each object to quantify the difference in object boundaries (Albrecht, Lang, and Hölbling Citation2010). The mean value and variance of these metrics can be used to measure the global accuracy (Schöpfer and Lang Citation2006).

2.5.2.4. Segmentation accuracy

For object-based image classification, segmentation is the procedure for creating the objects. When one validation object is divided into more than one object in the segmentation, over-segmenation occurs. When the segmentation produces objects that encompass more than one validation object, under-segmentation occurs. Over- and under-segmentation are the two major errors associated with segmentation. A large number of metrics for quantifying these two situations have been developed to represent segmentation accuracy. Over-segmentation metrics are mainly based on the ratio of overlapping area to validation object area. Under-segmentation metrics, on the other hand, are mainly based on the ratio of overlapping area to classified object area (Clinton et al. Citation2010; Persello and Bruzzone Citation2010). A comprehensive summary of these metrics can be found in Clinton et al. (Citation2010). Zhang et al. (Citation2015a) proposed precision as a global metric for under-segmentation assessment, by taking the ratio of all overlapping areas to all classified object areas; and recall for over-segmentation assessment, by taking the ratio of all overlapping areas to all validation object areas.

Combined metrics have also been proposed to consider both over- and under-segmentation in an integrated fashion (Zhang et al. Citation2015a; Clinton et al. Citation2010). Intuitive ways to combine these two metrics include the sum and root mean square (289–299; Möller et al. Citation2013). The F-measure has also been adopted for this combination (Zhang et al. Citation2015a).

2.5.3. Combination of geometric and thematic accuracies

Geometric and thematic accuracy measures are typically designed separately and are complementary to each other. Nevertheless, attempts have been made to combine these measures.

Lizarazo (Citation2014) proposed the STEP similarity matrix to comprehensively evaluate the accuracy of object-based image classification. The acronym STEP denotes shape, theme, edge, and position, respectively. The STEP matrix resembles the traditional CM, but shows the four similarities in each cell. Traditional area-based thematic CM, as well as shape CM, edge CM, and position CM, can be derived from the STEP similarity matrix.

Hernando et al. (Citation2012) combined the OFA concept with the traditional pixel-based error matrix concept to create an OFA matrix, to integrally evaluate the thematic accuracy of the classification result in an integrated approach.

Pattern-based accuracy assessment is another approach used to combine geometric and thematic accuracy assessments (Foody Citation2008). A traditional way of doing pattern-based assessment is to extract pattern indices (e.g. number of target objects, size of target objects) from both classified and validation maps, and to compare the similarity (Taubenbock et al. Citation2011; Potere et al. Citation2009). Dihkan et al. (Citation2013) proposed another pattern-based procedure to assess the accuracy of tea plantation mapping. A fuzzy local matching algorithm (Power, Simms, and White Citation2001) was introduced to measure the local similarity of patterns between the classified and validation maps. A map of the local similarity level is used to show the pattern-based accuracy both quantitatively and qualitatively. Furthermore, global accuracy is calculated from the weighted average of the local similarity measure for each object.

3. Incorporating spatial information in spectral unmixing

Spectral unmixing is the process of decomposing the measured spectrum of a mixed pixel into a set of pure spectral signatures called endmembers and their corresponding abundances, indicating the fractional area coverage of each endmember present in the pixel (Keshava and Mustard Citation2002). Early research on spectral unmixing dates back three decades (Adams, Smith, and Johnson Citation1986). Since then, many studies have exploited the spectral information inherent in remotely sensed imagery to analyse mixtures of distinct materials, treating each pixel as independent of its spatial neighbours. Not until recently has spatial information been accounted for in the development of spectral unmixing methods (Van Der Meer Citation1999; Rand and Keenan Citation2001; Roessner et al. Citation2001; Plaza et al. Citation2002; Maselli Citation2001). It has been demonstrated that the integration of both spatial and spectral information can lead to improvements in the unmixing results (Shi and Wang Citation2014). In the remainder of this section, we summarize the state-of-the-art spatial spectral unmixing methods according to the following three categories: 1) endmember extraction, 2) selection of endmember combinations, and 3) abundance estimation.

3.1. Endmember extraction

The technique of endmember extraction aims to determine endmember spectra from the remotely sensed image itself, taking advantage of the fact that image endmembers share the same spatial scale and atmospheric conditions as the image to be unmixed. A substantial number of endmember extraction algorithms have been proposed in the literature (Bioucas-Dias et al. Citation2012). Based on the theory of convex geometry, many algorithms search for the set of pixels that are located in the vertices of a simplex fit to the convex hull of the pixels in spectral space. However, these geometrically based algorithms utilize spectral properties of the remotely sensed imagery alone and tend to be susceptible to outlier pixels and noise. This problem can generally be alleviated by taking into account the spatial context of the pixels. The incorporation of spatial information in endmember extraction has culminated in the development of spatially oriented algorithms, the refinement of spectral-only algorithms, and the development of preprocessing modules. Spatially oriented algorithms are a fundamentally different approach to geometrically based algorithms – the former rely on detecting pure pixels or pure neighbourhoods in the spatial context, rather than investigating spectral extremeness of pixels in the spectral space. The refined spectral-only algorithms make use of spatial information to improve the performance of geometrically based algorithms; the primary structures of the original spectral-only algorithms are not changed substantially. The preprocessing methods prior to endmember extraction have a unique modularity; no modification of existing endmember extraction algorithms is needed to apply the preprocessing methods. Extracted endmembers that account for the spatial context of candidate pixels may be more spatially representative and less susceptible to outlier pixels.

Spatially oriented endmember extraction algorithms provide a way to assess the purity of a pixel or a spatial neighbourhood without resorting to geometrically based algorithms (Torres-Madronero and Velez-Reyes Citation2014; Plaza et al. Citation2002; Mei et al. Citation2010). The automated morphological endmember extraction (AMEE) algorithm proposed by Plaza et al. (Citation2002) is the earliest attempt in this direction. This algorithm extends conventional greyscale morphology to multi-channel imagery and uses two morphological operations, i.e. dilation and erosion, to identify the respective pixels with highest and lowest purity in a sliding window. A morphological eccentricity index (MEI) value, defined as the spectral angle distance between the purest pixel and the most mixed pixel, is assigned to the purest pixel in the window. Endmembers are obtained by applying a threshold to the MEI image. More recently, Mei et al. (Citation2010) developed the spatial purity-based endmember extraction (SPEE) algorithm to detect pure spatial neighbourhoods. The spatial neighbourhood purity index is determined by either an intensity-level measurement used in the original intensity domain or feature-level measurement used in the transformation domain. A threshold is then applied to obtain endmember candidates. A graph-based spatial refinement process is adopted to reduce the number of endmember candidates. Pure spatial neighbourhood can also be identified using multi-scale representation of remotely sensed imagery (Torres-Madronero and Velez-Reyes Citation2014). A series of smoothed images are constructed and endmember spectra are extracted from the multi-grid structure. The smoothed spectra are equivalent to the average of spectrally similar and spatially adjacent pixels.

In addition, a collection of endmember extraction algorithms that incorporate spatial information has been proposed to refine spectral-only algorithms (Li and Zhang Citation2011a; Rogge et al. Citation2007; Zhang, Rivard, and Rogge Citation2008; Xu, Du, and Zhang Citation2014; Xu, Zhang, and Du Citation2015). These either average spectrally similar and spatially adjacent endmember candidates produced by spectral-only algorithms, or select endmember candidates that are present in homogenous neighbourhoods. The spatial-spectral endmember extraction (SSEE) algorithm (Rogge et al. Citation2007) is an improvement on pixel purity index (PPI) (Boardman Citation1993; Boardman, Kruse, and Green Citation1995). A set of local eigenvectors generated by applying singular value decomposition (SVD) to non-overlapping image subsets is used as projection vectors for PPI. The obtained endmember candidates are updated by a spatial averaging process to average spectrally similar endmember candidates within a sliding window. Zhang, Rivard, and Rogge (Citation2008) improved orthogonal subspace projection (OSP) (Harsanyi and Chang Citation1994) and vertex component analysis (VCA) (Nascimento and Dias Citation2005) by averaging the endmember candidates that met both the spectral similarity and spatial adjacency criteria. Li and Zhang (Citation2011a) refined the iterative error analysis (IEA) approach (Neville et al. Citation1999) by identifying endmember candidates that had sufficient spectrally similar pixels in a local window. Xu, Du, and Zhang (Citation2014) developed the abundance-constrained endmember extraction (ACEE) algorithm and further incorporated spatial information in ACEE. A unique feature of the proposed spatial-spectral information-based ACEE (SSACEE) is that local homogeneity of endmember candidates is determined using estimated abundances instead of pixel spectra. Xu, Zhang, and Du (Citation2015) used PPI as a coarse screening. In each partitioned image subset, endmember candidates identified by PPI were further evaluated according to local homogeneity criteria in order to obtain an endmember set.

In contrast to the incorporation of spatial information in endmember extraction algorithms, a number of studies consider spatial information through a preprocessing step prior to endmember extraction. This preprocessing module can be combined with any endmember extraction algorithm. In the spatial preprocessing (SPP) algorithm (Zortea and Plaza Citation2009), a spatially derived weighting factor is used to adjust the original spectral signature on a per-pixel basis so that pixels located in spatially homogenous regions will engage a lower number of adjustments than those in spatially heterogeneous regions. As such, when a geometrically based algorithm is applied to the adjusted image, pixels located in spatially homogenous regions are more likely to be identified as endmembers. Many preprocessing methods adopt segmentation or clustering techniques to partition remotely sensed imagery into homogenous regions and use the mean spectra of each region for endmember extraction algorithms (Thompson et al. Citation2010; Martin and Plaza Citation2011; Zhang et al. Citation2014a). The spatial-spectral preprocessing (SSPP) algorithm (Martin and Plaza Citation2012) provides a more integrated framework to combine both spatial homogeneity and spectral purity at the preprocessing level. Within each region generated by unsupervised clustering, a subset of spatially homogenous and spectrally pure pixels is identified. Spatial and spectral information can also be combined to develop local neighbourhood weights (Liu et al. Citation2012). A scene is divided into homogenous and transition regions through weight thresholding. Only pixels that fall in homogenous regions are analysed by endmember extraction algorithms. Several preprocessing methods have recently been developed to reduce the size of the original data set while retaining the accuracy of endmember extraction (Rogge et al. Citation2012; Lopez et al. Citation2013; Beauchemin Citation2014). Rogge et al. (Citation2012) applied local endmember extraction to non-overlapping image subsets to constitute a reduced set of pixels. Lopez et al. (Citation2013) selected pixels that were at the spatial edges and spectral extremes of the image. Beauchemin (Citation2014) exploited locally identified pixels that were spectrally similar to the mean spectra of non-overlapping image subsets to define a simplex, and filtered out the pixels located inside this simplex.

A potential drawback of spatial spectral algorithms for endmember extraction is that these algorithms tend to neglect anomalous endmembers (Zortea and Plaza Citation2009). To overcome this problem, anomaly detection techniques can be integrated with spatial spectral algorithms to extract both homogenous and anomalous endmembers (Mei et al. Citation2011; Erturk et al. Citation2014).

3.2. Selection of endmember combinations

Most spectral unmixing studies assume that the set of endmembers is invariant across the entire scene. However, due to endmember variability arising from changing illumination conditions and physical differences in broadly defined endmember classes, it is more realistic to use multiple endmember spectra to represent the same class (Somers et al. Citation2011; Zare and Ho Citation2014). Moreover, the actual number of endmembers contained in a single pixel is usually much lower than the number of endmembers present across the entire scene (Roberts et al. Citation1998). In order to address these two issues, many spectral-only methods have been developed to allow the number and type of endmembers as well as their corresponding spectral signatures to vary on a per-pixel basis (Somers et al. Citation2011), e.g. the well-known multiple endmember spectral mixture analysis (MESMA) (Roberts et al. Citation1998). However, these methods usually suffer from spectral confusion and computational complexity induced by numerous endmember combinations. Considering that nearby pixels are likely to share similar endmember combinations, the incorporation of spatial information in the selection of endmember combinations can alleviate spectral confusion and improve computational efficiency. The spatial spectral methods available for selection of endmember combinations can be categorized into two groups: per-pixel methods that allow each pixel to have a specific set of endmembers, and per-field methods that select a distinct endmember combination for each homogenous field partitioned from remotely sensed imagery.

When pure pixels or image endmembers are widely distributed across an image, a per-pixel strategy is usually applied to select for a mixed pixel a set of endmembers that are spatially close to the pixel. Spectrally pure pixels can be identified by supervised classification with rigorous criteria (Roessner et al. Citation2001; Deng and Wu. Citation2013; Wu, Deng, and Jia Citation2014; King and Younan Citation2006), manual interpretation (Johnson, Tateishi, and Kobayashi Citation2012; Zhong, Wang, and Wu Citation2015), maximum abundances of global endmembers (Cui, Li, and Zhao Citation2014), histogram thresholding (Ma et al. Citation2014), or purity determination algorithms (Mei, Du, and He Citation2015). For a mixed pixel located at the centre of a predefined spatial neighbourhood, the optimal per-pixel endmember set can be selected by an iterative unmixing procedure similar to MESMA (Roessner et al. Citation2001; Wu, Deng, and Jia Citation2014), an endmember scoring scheme (King and Younan Citation2006; Mei, Du, and He Citation2015), or an endmember synthesis process via spatial averaging (Deng and Wu. Citation2013; Ma et al. Citation2014; Zhong, Wang, and Wu Citation2015). It is worth noting that Johnson, Tateishi, and Kobayashi (Citation2012) used a more sophisticated spatial interpolator, i.e. ordinary kriging (Van Der Meer Citation2012), to synthesize the vegetation indices of bare soil (and/or dead vegetation) and full green vegetation cover, and produced more accurate estimates of fractional green vegetation cover than through the use of inverse distance weighting (IDW) for synthesis. In addition to interpolation, extrapolation has been used to synthesize endmember spectra for coarse-resolution images where pure pixels are absent (Maselli Citation2001; Zhang et al. Citation2015b). With known endmember fractions of training samples, multivariate linear regression models are constructed using the fractions as independent variables and the spectra of the sample pixels as dependent variables. Endmember spectra are then synthesized by extrapolating the relevant regression model to 100% cover fraction for each endmember. In order to incorporate endmember variability, Maselli (Citation2001) used locally calibrated regression statistics, including mean vectors and variance/covariance matrices, to construct regression models. Zhang et al. (Citation2015b) adopted geographically weighted regression (GWR) (Brunsdon, Fotheringham, and Charlton Citation1996) to derive synthetic endmember spectra. Both methods can produce spatially variable endmember signatures. Furthermore, an improvement on MESMA has been made through a hierarchical approach (Franke et al. Citation2009). Classification results from simpler levels are used as spatial constraints for selection of endmember combinations at more detailed levels. It should be noted that all the aforementioned per-pixel methods represent endmembers as a discrete set. Nonetheless, endmembers can also be represented using a continuous distribution. A recent study exploited the beta compositional model (BCM) for spectral unmixing (Du et al. Citation2014). Each pixel was considered to follow the distribution of a convex combination of beta-distributed endmembers, which was also approximated by beta distribution. In the proposed BCM-spatial algorithm, K-nearest spatial-spectral neighbours were identified and used to estimate the statistical moments of the beta distribution for each pixel. These statistical moments were then input to the objective function or the likelihood function to estimate the abundances.

The per-field methods for selection of endmember combinations generally partition a heterogeneous scene into a number of homogeneous fields and assign each field a small set of endmembers that are present in the field. A variety of image partitioning approaches have been adopted, including Markov random fields (MRFs) (Rand and Keenan Citation2001), iterative self-organizing data analysis techniques A (ISODATA) (Shoshany and Svoray Citation2002), unmixing results (García‐Haro, Sommer, and Kemper Citation2005), local tiles (Canham et al. Citation2011; Goenaga et al. Citation2013), iterative fuzzy clustering (Zare et al. Citation2013), image classification (Zhang et al. Citation2014c), and multi-resolution image segmentation (Zhang et al. Citation2014a; Li et al. Citation2015). The image partitions so produced are commonly grouped to form homogenous fields that represent different land-cover types according to prior knowledge of the scene (Rand and Keenan Citation2001; Shoshany and Svoray Citation2002; García‐Haro, Sommer, and Kemper Citation2005; Zhang et al. Citation2014c). This image stratification process is also suggested by a comparative study for estimating fractional green vegetation cover (Xiao and Moody Citation2005). Endmembers in each field are then selected by a manual process (Rand and Keenan Citation2001; Shoshany and Svoray Citation2002; García‐Haro, Sommer, and Kemper Citation2005; Zhang et al. Citation2014c; Li et al. Citation2015) or automatic endmember extraction algorithms (Canham et al. Citation2011; Goenaga et al. Citation2013; Zhang et al. Citation2014a). Zare et al. (Citation2013) proposed a piecewise convex multiple-model endmember detection (PCOMMEND) algorithm to partition an image into multiple fields constrained by a family of simplices in spectral space. A joint optimization procedure was conducted to estimate endmember spectra, fractional abundances, and fuzzy membership values assigned to each pixel denoting the field to which it belonged. An improved algorithm was developed to add spatial constraints on fuzzy membership values in order to guide neighbouring pixels to share the same set of endmembers (Zare et al. Citation2010).

3.3. Abundance estimation

Fractional abundances of endmembers present in mixed pixels are the primary product of spectral unmixing. The spectrum of a pixel is often modelled as a linear combination of endmember spectra weighted by their corresponding abundances (Settle and Drake Citation1993; Shimabukuro and Smith Citation1991), and the abundances can be derived using the least squares method to minimize the mean square error between the actual spectrum and the reconstructed spectrum (Heinz and Chang Citation2001). In spite of its mathematic simplicity, this method estimates abundances in a pixel-by-pixel fashion and the abundances of each pixel are derived independently of its neighbouring pixels. The incorporation of spatial information, nonetheless, can improve the accuracy of abundance estimation and encourage derived abundances to be more spatially consistent.

An early exploratory work was conducted by Van Der Meer (Citation1999). A series of spatial criteria defined on the RMSE for each pixel were employed for an iterative spectral unmixing (ISU) method with the goal of removing spatial structure remaining in the residuals. However, spatial criteria were not included in an objective function to be optimized. In the spectral and spatial complexity blind source separation (SSCBSS) method (Jia and Qian Citation2007), the abundance in a pixel was predicted by the weighted average of the abundances in its neighbouring pixels and the objective function was constructed to maximize the spatial predictability of the abundances. A MRF was also exploited for spectral unmixing (Eches, Dobigeon, and Tourneret Citation2011). In the proposed method, an implicit image classification was conducted to partition the image into homogenous regions where the statistical moments of fractional abundances were unchanged. A hierarchical Bayesian model was then adopted to infer the distributions of the abundances and class labels. In addition, local homogeneity was exploited to select landmark points for nonlinear manifold-based spectral unmixing (Bottai et al. Citation2013). Recently, image partitioning was integrated with a reweighted λ1 framework in a spatially weighted sparse coding (SWSC) algorithm (Soltani-Farani and Rabiee Citation2015).

Arguably, the most popular spatial spectral strategy for abundance estimation makes use of a term of spatial regularization that is involved in the optimization procedure to promote the smoothness of estimated abundances. Examples of regularizers include spatial statistics (Song, Jiang, and Rui Citation2010), fuzzy local information (Zare Citation2011), spatial spectral coherence (Castrodad et al. Citation2011), total variation (Chen, Richard, and Honeine Citation2014; Iordache, Bioucas-Dias, and Plaza Citation2012), λ 2-norm (Liu et al. Citation2012; Zhu et al. Citation2014), and others (Jia and Qian Citation2009; Liu et al. Citation2011). Spatial regularization is especially important for sparse unmixing, which aims to find the optimal subset of endmember spectra in a spectral library through sparsity-inducing regularization (Iordache, Bioucas-Dias, and Plaza Citation2011). Since the number of endmembers present in a single pixel is relatively limited, the true abundances associated with a large spectral library are inevitably comprised of many zeros. Therefore, the abundance maps depict pronounced spatial autocorrelation that can be taken into account in the optimization procedure through spatial regularization.

4. Summary and future perspectives

Incorporating spatial information in spectral-based image classification is an important aspect many remote-sensing studies. In this paper, we reviewed and summarized recent advances in various techniques related to this topic and grouped them into five sets: pre-classification methods, sample selection, classifiers, post-classification, and accuracy assessment. Before remote-sensing images are fed to various classifiers, spatial information in most cases is incorporated as additional bands in the original spectral images. GLCM, as a texture measurement, is probably the most widely and frequently adopted method in this case. Recent advancements in other methods to derive texture information, such as fractal dimension, geostatistics, and Gabor and wavelet filters, or image morphological analysis, have also shown their potential to extract spatial information at the pre-classification stage. However, the performance of these different methods may be dependent on the specific application and a comprehensive comparison of these methods has yet to be made.

By including spatial information in sample selection, the sample unit, sample size, and sample design are all affected. A number of studies have incorporated geostatistics into sampling to help determine the sample unit size or even sampling design. However, these studies have mainly focused on interpolation, but not classification. To some extent, to predict the class labels from training samples is also a type of interpolation. Therefore, the conclusions and analytical methods can be borrowed for sample selection in image classification.

GEOBIA, as a new and evolving paradigm in remote sensing and GIScience, provides a spatial perspective for understanding image objects and geographic-based intelligence. In the future, we need to build a more comprehensive geographic knowledge-sharing system with GEOBIA databases, to support collective intelligence and knowledge accumulation. Additionally, object-oriented data models need to be integrated into GEOBIA to more effectively explain the natural and anthropogenic phenomenon with image objects. Two contextual classifiers (SVM and MRF) are primarily reviewed in this study to exploit the spatial information among neighbouring pixels. The introduction of composite kernels in SVM offers valuable opportunities to exploit the spatial information in classifiers. To better take advantage of the diverse spatial information (e.g. shape, size, and homogeneity), more sophisticated spatial kernels need to be explored. MRFs exhibit considerable capacity for incorporating multiple information sources (e.g. multi-scale, multi-bands, multi-temporal) in land-cover mapping. Investigating various information sources in a comprehensive framework is thus needed to advance this classifier.

The emphasis on spatial information has encouraged the development of spatially related accuracy assessment for image classification. Various metrics have been developed for the assessment of both geometric and thematic accuracy of image classification results. Among these, some are conceptually identical and technically similar. A thorough compilation of existing accuracy metrics is needed. In addition, although some interesting topics have been raised, spatial-thematic accuracy assessment is still underdeveloped.

The incorporation of spatial information has also achieved great success for improving the performance of spectral unmixing. In this paper, we summarized the available spatial spectral unmixing methods in the three aspects of spectral unmixing: endmember extraction, selection of endmember combinations, and abundance estimation. Compared with the extent to which spatial information is incorporated in image classification, the area of spatial spectral unmixing is less established and requires more attention from the remote-sensing research community. With this review we hope to trigger new studies to expedite the incorporation of spatial information in spectral unmixing.

Disclosure statement

No potential conflict of interest was reported by the authors.

References

  • Adams, J. B., M. O. Smith, and P. E. Johnson. 1986. “Spectral Mixture Modeling - a New Analysis of Rock and Soil Types at the Viking Lander-1 Site.” Journal of Geophysical Research-Solid Earth and Planets 91 (B8): 8098–8112. doi:10.1029/JB091iB08p08098.
  • Albrecht, F., S. Lang, and D. Hölbling. 2010. “Spatial Accuracy Assessment of Object Boundaries for Object-Based Image Analysis.” International Archives of the Photogrammetry Remote Sensing and Spatial Information Sciences 38 (4): C7.
  • Andres, S., D. Arvor, and C. Pierkot. 2012. “Towards an Ontological Approach for Classifying Remote Sensing Images.” Paper presented at the 2012 Eighth International Conference on Signal Image Technology and Internet Based Systems (SITIS), Naples.
  • Angel, S., J. Parent, and D. L. Civco. 2010. “Ten Compactness Properties of Circles: Measuring Shape in Geography.” The Canadian Geographer 54 (4): 441–461.
  • Angelo, N. P., and V. Haertel. 2003. “On the Application of Gabor Filtering in Supervised Image Classification.” International Journal of Remote Sensing 24 (10): 2167–2189. doi:10.1080/01431160210163146.
  • Atkinson, P. M. 2004. “Spatially Weighted Supervised Classification for Remote Sensing.” International Journal of Applied Earth Observation and Geoinformation 5 (4): 277–291. doi:10.1016/j.jag.2004.07.006.
  • Atkinson, P. M., and D. K. Naser. 2010. “A Geostatistically Weighted k‐NN Classifier for Remotely Sensed Imagery.” Geographical Analysis 42 (2): 204–225. doi:10.1111/j.1538-4632.2010.00790.x.
  • Atkinson, P. M., and P. Lewis. 2000. “Geostatistical Classification for Remote Sensing: An Introduction.” Computers & Geosciences 26 (4): 361–371. doi:10.1016/S0098-3004(99)00117-X.
  • Baatz, M., and S. Arno. 2000. “Multiresolution Segmentation: An Optimization Approach for High Quality Multi-Scale Image Segmentation.” Angewandte Geographische InformationsverarbeitungXII: 12–23.
  • Baatz, M., U. Benz, S. Dehghani, M. Heynen, A. Höltje, P. Hofmann, I. Lingenfelder, M. Mimler, M. Sohlbach, and M. Weber. 2000. Ecognition User Guide.Munich: Definiens.
  • Baatz, M., and A. Schäpe. 1999. “Object-Oriented and Multi-Scale Image Analysis in Semantic Networks.” Paper Presented at the 2nd International Symposium: Operationalization of Remote Sensing, ITC, Enschede.
  • Balaguer, A., L. A. Ruiz, T. Hermosilla, and J. A. Recio. 2010. “Definition of a Comprehensive Set of Texture Semivariogram Features and Their Evaluation for Object-Oriented Image Classification.” Computers & Geosciences 36 (2): 231–240. doi:10.1016/j.cageo.2009.05.003.
  • Baraldi, A., L. Boschetti, and M. L.Humber. 2014. “Probability Sampling Protocol for Thematic and Spatial Quality Assessment of Classification Maps Generated From Spaceborne/Airborne Very High Resolution Images.” IEEE Transactions on Geoscience and Remote Sensing 52 (1): 701–760. doi:10.1109/TGRS.2013.2243739.
  • Barnsley, M. J., and S. L. Barr. 1996. “Inferring Urban Land Use from Satellite Sensor Images Using Kernel-Based Spatial Reclassification.” Photogrammetric Engineering and Remote Sensing 62 (8): 949–958.
  • Bau, T. C., S. Sarkar, and G. Healey. 2010. “Hyperspectral Region Classification Using a Three-Dimensional Gabor Filterbank.” IEEE Transactions on Geoscience and Remote Sensing 48 (9): 3457–3464. doi:10.1109/TGRS.2010.2046494.
  • Beauchemin, M. 2014. “A Method Based on Spatial and Spectral Information to Reduce the Solution Space in Endmember Extraction Algorithms.” Remote Sensing Letters 5 (5): 471–480. doi:10.1080/2150704x.2014.920549.
  • Beaulieu, J.-M., and M. Goldberg. 1989. “Hierarchy in Picture Segmentation: A Stepwise Optimization Approach.” IEEE Transactions on Pattern Analysis and Machine Intelligence 11 (2): 150–163. doi:10.1109/34.16711.
  • Bellens, R., S. Gautama, L. Martinez-Fonte, W. Philips, J. C.-W.Chan, and F. Canters. 2008. “Improved Classification of VHR Images of Urban Areas Using Directional Morphological Profiles.” IEEE Transactions on Geoscience and Remote Sensing 46 (10): 2803–2813. doi:10.1109/TGRS.2008.2000628.
  • Benediktsson, J. A., J. A.Palmason, and J. R.Sveinsson. 2005. “Classification of Hyperspectral Data from Urban Areas Based on Extended Morphological Profiles.” IEEE Transactions on Geoscience and Remote Sensing 43 (3): 480–491. doi:10.1109/TGRS.2004.842478.
  • Benz, U. C., P. Hofmann, G. Willhauck, I. Lingenfelder, and M. Heynen. 2004. “Multi-Resolution, Object-Oriented Fuzzy Analysis of Remote Sensing Data for GIS-Ready Information.” ISPRS Journal of Photogrammetry and Remote Sensing 58 (3–4): 239–258. doi:10.1016/j.isprsjprs.2003.10.002.
  • Berberoglu, S., P. J. Curran, C. D. Lloyd, and P. M. Atkinson. 2007. “Texture Classification of Mediterranean Land Cover.” International Journal of Applied Earth Observation and Geoinformation 9 (3): 322–334. doi:10.1016/j.jag.2006.11.004.
  • Besag, J. 1974. “Spatial Interaction and the Statistical Analysis of Lattice Systems.” Journal of the Royal Statistical Society Series B (Methodological) 36 (2):192–236.
  • Bioucas-Dias, J. M., A. Plaza, N. Dobigeon, M. Parente, Q. Du, P. Gader, and J. Chanussot. 2012. “Hyperspectral Unmixing Overview: Geometrical, Statistical, and Sparse Regression-Based Approaches.” IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing 5 (2): 354–379. doi:10.1109/jstars.2012.2194696.
  • Bioucas-Dias, J. M., A. Plaza, G. Camps-Valls, P. Scheunders, N. A. S. S. E. R. M. Nasrabadi, and J. Chanussot. 2013. “Hyperspectral Remote Sensing Data Analysis and Future Challenges.” IEEE Geoscience and Remote Sensing Magazine 1 (2): 6–36. doi:10.1109/MGRS.2013.2244672.
  • Blaschke, T. 2010. “Object Based Image Analysis for Remote Sensing.” ISPRS Journal of Photogrammetry and Remote Sensing 65 (1): 2–16. doi:10.1016/j.isprsjprs.2009.06.004.
  • Blaschke, T., C. Burnett, and A. Pekkarinen. 2004. “Image Segmentation Methods for Object-Based Analysis and Classification.” In Remote Sensing Image Analysis: Including the Spatial Domain, edited by S. M. de Jong and F. D. van der Meer, 211–236. Dordrecht: Kluwer Academic Publishers.
  • Blaschke, T., G. J. Hay, M. Kelly, S. Lang, P. Hofmann, E. Addink, R. Q. Feitosa, F. van der Meer, H. van der Werff, F. van Coillie, and D. Tiede. 2014. “Geographic Object-Based Image Analysis – Towards a New Paradigm.” ISPRS Journal of Photogrammetry and Remote Sensing 87: 180–191. doi:10.1016/j.isprsjprs.2013.09.014.
  • Blaschke, T., S. Lang, and G. Hay. 2008. Object-Based Image Analysis: Spatial Concepts for Knowledge-Driven Remote Sensing Applications. Heidelberg: Springer-verlag.
  • Blaschke, T., and J. Strobl. 2001. “What’s Wrong with Pixels? Some Recent Developments Interfacing Remote Sensing and GIS.” GeoBIT/GIS 6 (1): 12–17.
  • Boardman, J. W. 1993. “Automating Spectral Unmixing of AVIRIS Data using Convex Geometry Concepts.” Paper presented at the Summaries of the Fourth Annual JPL Airborne Geoscience Workshop, Washington, DC.
  • Boardman, J. W., K. A. Kruse, and R. O. Green. 1995. “Mapping Target Signatures via Partial Unmixing of AVIRIS Data.” Paper presented at the Summaries of the Fifth Annual JPL Airborne Geoscience Workshop, Pasadena, CA.
  • Bottai, L., L. Arcidiaco, M. Chiesi, and F. Maselli. 2013. “Application of a Single-Tree Identification Algorithm to Lidar Data for the Simulation of Stem Volume Current Annual Increment.” Journal of Applied Remote Sensing 7: 073699. doi:10.1117/1.JRS.7.073699.
  • Boykov, Y., and G. Funka-Lea. 2006. “Graph Cuts and Efficient ND Image Segmentation.” International Journal of Computer Vision 70 (2): 109–131. doi:10.1007/s11263-006-7934-5.
  • Boykov, Y., O. Veksler, and R. Zabih. 2001. “Fast Approximate Energy Minimization via Graph Cuts.” IEEE Transactions on Pattern Analysis and Machine Intelligence 23 (11): 1222–1239. doi:10.1109/34.969114.
  • Brunsdon, C., A. S. Fotheringham, and M. E. Charlton. 1996. “Geographically Weighted Regression: A Method for Exploring Spatial Nonstationarity.” Geographical Analysis 28 (4): 281–298. doi:10.1111/gean.1996.28.issue-4.
  • Burnett, C., and T. Blaschke. 2003. “A Multi-Scale Segmentation/Object Relationship Modelling Methodology for Landscape Analysis.” Ecological Modelling 168 (3): 233–249. doi:10.1016/S0304-3800(03)00139-X.
  • Campbell, J. B. 1981. “Spatial Correlation Effects upon Accuracy of Supervised Classification of Land Cover.” Photogrammetric Engineering & Remote Sensing 47 (3): 355–363.
  • Camps-Valls, G., L. Gomez-Chova, J. Munoz-Mari, J. Vila-Frances, and J. Calpe-Maravilla. 2006. “Composite Kernels for Hyperspectral Image Classification.” IEEE Geoscience and Remote Sensing Letters 3 (1): 93–97. doi:10.1109/LGRS.2005.857031.
  • Camps-Valls, G., N. Shervashidze, and K. M.Borgwardt. 2010. “Spatio-Spectral Remote Sensing Image Classification with Graph Kernels.” IEEE Geoscience and Remote Sensing Letters 7 (4): 741–745. doi:10.1109/LGRS.2010.2046618.
  • Camps-Valls, G., D. Tuia, L. Bruzzone, and J. Atli Benediktsson. 2014. “Advances in Hyperspectral Image Classification: Earth Monitoring with Statistical Learning Methods.” IEEE Signal Processing Magazine 31 (1): 45–54. doi:10.1109/MSP.2013.2279179.
  • Canham, K., A. Schlamm, A. Ziemann, B. Basener, and D. Messinger. 2011. “Spatially Adaptive Hyperspectral Unmixing.” IEEE Transactions on Geoscience and Remote Sensing 49 (11): 4248–4262. doi:10.1109/TGRS.2011.2169680.
  • Carneiro, J. T., and M. J. Pereira. 2014. “Beyond the Confusion Matrix: Geostatistical Error Assessment for Landsat Landcover Maps of the Portuguese Landscape.” Paper presented at the 2014 IEEE International on Geoscience and Remote Sensing Symposium (IGARSS), Quebec City, QC.
  • Carr, J. R., and F. P.De Miranda. 1998. “The Semivariogram in Comparison to the Co-Occurrence Matrix for Classification of Image Texture.” IEEE Transactions on Geoscience and Remote Sensing 36 (6): 1945–1952. doi:10.1109/36.729366.
  • Castilla, G., A. Hernando, C. H. Zhang, and G. J. McDermid. 2014. “The Impact of Object Size on the Thematic Accuracy of Landcover Maps.” International Journal of Remote Sensing 35 (3): 1029–1037. doi:10.1080/01431161.2013.875630.
  • Castilla, G., G. G.Hay, and J. R.Ruiz-Gallardo. 2008. “Size-constrained Region Merging (SCRM).” Photogrammetric Engineering & Remote Sensing 74 (4): 409–419. doi:10.14358/PERS.74.4.409.
  • Castrodad, A., Z. M. Xing, J. B. Greer, E. Bosch, L. Carin, and G. Sapiro. 2011. “Learning Discriminative Sparse Representations for Modeling, Source Separation, and Mapping of Hyperspectral Imagery.” IEEE Transactions on Geoscience and Remote Sensing 49 (11): 4263–4281. doi:10.1109/tgrs.2011.2163822.
  • Celeux, G., and G. Govaert. 1992. “A Classification EM Algorithm for Clustering and Two Stochastic Versions.” Computational Statistics & Data Analysis 14 (3): 315–332. doi:10.1016/0167-9473(92)90042-E.
  • Chang, C.-C., and C.-J. Lin. 2011. “LIBSVM: A Library for Support Vector Machines.” ACM Transactions on Intelligent Systems and Technology (TIST)2 (3): 27.
  • Chen, D., D. A. Stow, and P. Gong. 2004. “Examining the Effect of Spatial Resolution and Texture Window Size on Classification Accuracy: An Urban Environment Case.” International Journal of Remote Sensing 25 (11): 2177–2192. doi:10.1080/01431160310001618464.
  • Chen, D., and D. Stow. 2002. “The Effect of Training Strategies on Supervised Classification at Different Spatial Resolutions.” Photogrammetric Engineering and Remote Sensing 68 (11): 1155–1162.
  • Chen, J., C. Richard, and P. Honeine. 2014. “Nonlinear Estimation of Material Abundances in Hyperspectral Images With l(1)-Norm Spatial Regularization.” IEEE Transactions on Geoscience and Remote Sensing 52 (5): 2654–2665. doi:10.1109/tgrs.2013.2264392.
  • Chen, Y., N. M.Nasrabadi, and T. D.Tran. 2011. “Hyperspectral Image Classification Using Dictionary-Based Sparse Representation.” IEEE Transactions on Geoscience and Remote Sensing 49 (10): 3973–3985. doi:10.1109/TGRS.2011.2129595.
  • Chen, Y., N. M.Nasrabadi, and T. D.Tran. 2013. “Hyperspectral Image Classification via Kernel Sparse Representation.” IEEE Transactions on Geoscience and Remote Sensing 51 (1): 217–231. doi:10.1109/TGRS.2012.2201730.
  • Chen, Y., K. Qin, S. Gan, and T. Wu. 2014. “Structural Feature Modeling of High-Resolution Remote Sensing Images Using Directional Spatial Correlation.” IEEE Geoscience and Remote Sensing Letters 11 (10): 1727–1731. doi:10.1109/LGRS.2014.2306972.
  • Clausi, D. A., and Y. Zhao. 2002. “Rapid Extraction of Image Texture by Co-Occurrence Using a Hybrid Data Structure.” Computers & Geosciences 28 (6): 763–774. doi:10.1016/S0098-3004(01)00108-X.
  • Clausi, D. A., and Y. Zhao. 2003. “Grey Level Co-Occurrence Integrated Algorithm (GLCIA): A Superior Computational Method to Rapidly Determine Co-Occurrence Probability Texture Features.” Computers & Geosciences 29 (7): 837–850. doi:10.1016/s0098-3004(03)00089-x.
  • Clinton, N., A. Holt, J. Scarborough, L. Yan, and P. Gong. 2010. “Accuracy Assessment Measures for Object-based Image Segmentation Goodness.” Photogrammetric Engineering & Remote Sensing 76 (3): 289–299. doi:10.14358/PERS.76.3.289.
  • Conchedda, G., L. Durieux, and P. Mayaux. 2008. “An Object-Based Method for Mapping and Change Analysis in Mangrove Ecosystems.” ISPRS Journal of Photogrammetry and Remote Sensing 63 (5): 578–589. doi:10.1016/j.isprsjprs.2008.04.002.
  • Congalton, R. G. 1991. “A Review of Assessing the Accuracy of Classifications of Remotely Sensed Data.” Remote Sensing of Environment 37 (1): 35–46. doi:10.1016/0034-4257(91)90048-B.
  • Congalton, R. G., J. Y. Gu, K. Yadav, P. Thenkabail, and M. Ozdogan. 2014. “Global Land Cover Mapping: A Review and Uncertainty Analysis.” Remote Sensing 6 (12): 12070–12093. doi:10.3390/rs61212070.
  • Congalton, R. G., and K. Green. 2009. Assessing the Accuracy of Remotely Sensed Data: Principles and Practices. 2nd ed. Boca Raton, FL: CRC press.
  • Cortes, C., and V. Vapnik. 1995. “Support-Vector Networks.” Machine Learning 20 (3): 273–297. doi:10.1007/BF00994018.
  • Cristianini, N., and S.-T. John. 2000. An Introduction to Support Vector Machines and Other Kernel-Based Learning Methods. Cambridge: Cambridge university press.
  • Cui, J. T., X. R. Li, and L. Y. Zhao. 2014. “Nonlinear Spectral Mixture Analysis by Determining Per-Pixel Endmember Sets.” IEEE Geoscience and Remote Sensing Letters 11 (8): 1404–1408. doi:10.1109/lgrs.2013.2294181.
  • Curran, P. J., and P. M.Atkinson. 1998. “Geostatistics and Remote Sensing.” Progress in Physical Geography 22 (1): 61–78. doi:10.1177/030913339802200103.
  • Dalla Mura, M., J. A.Benediktsson, B. Waske, and L. Bruzzone. 2010. “Morphological Attribute Profiles for the Analysis of Very High Resolution Images.” IEEE Transactions on Geoscience and Remote Sensing 48 (10): 3747–3762. doi:10.1109/TGRS.2010.2048116.
  • Das, S. K., and R. Singh. 2009. “Performance of Kriging-Based Soft Classification on Wifs/IRS-1D Image Using Ground Hyperspectral Signatures.” IEEE Geoscience and Remote Sensing Letters 6 (3): 453–457. doi:10.1109/LGRS.2009.2016988.
  • de Bruin, S. 2000. “Predicting the Areal Extent of Land-Cover Types Using Classified Imagery and Geostatistics.” Remote Sensing of Environment 74 (3): 387–396. doi:10.1016/S0034-4257(00)00132-2.
  • De Jong, S. M., and P. A. Burrough. 1995. “A Fractal Approach to the Classificatoin of Mediterranean Vegetation Types in Remotely Sensed Images.” Photogrammetric Engineering & Remote Sensing 61 (8): 1041–1053.
  • Deng, C., and C. Wu. 2013. “A Spatially Adaptive Spectral Mixture Analysis for Mapping Subpixel Urban Impervious Surface Distribution.” Remote Sensing of Environment 133: 62–70. doi:10.1016/j.rse.2013.02.005.
  • Desclée, B., P. Bogaert, and P. Defourny. 2006. “Forest Change Detection by Statistical Object-Based Method.” Remote Sensing of Environment 102 (1–2): 1–11. doi:10.1016/j.rse.2006.01.013.
  • Dihkan, M., N. Guneroglu, F. Karsli, and A. Guneroglu. 2013. “Remote Sensing of Tea Plantations Using an SVM Classifier and Pattern-Based Accuracy Assessment Technique.” International Journal of Remote Sensing 34 (23): 8549–8565. doi:10.1080/01431161.2013.845317.
  • Du, X. X., A. Zare, P. Gader, and D. Dranishnikov. 2014. “Spatial and Spectral Unmixing Using the Beta Compositional Model.” IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing 7 (6): 1994–2003. doi:10.1109/jstars.2014.2330347.
  • Dubes, R. C., and A. K.Jain. 1989. “Random Field Models in Image Analysis.” Journal of Applied Statistics 16 (2): 131–164. doi:10.1080/02664768900000014.
  • Eches, O., N. Dobigeon, and J.-Y.Tourneret. 2011. “Enhancing Hyperspectral Image Unmixing With Spatial Correlations.” IEEE Transactions on Geoscience and Remote Sensing 49 (11): 4239–4247. doi:10.1109/tgrs.2011.2140119.
  • Emerson, C. W., N. S.-N.Lam, and D. A.Quattrochi. 2005. “A Comparison of Local Variance, Fractal Dimension, and Moran’s I as Aids to Multispectral Image Classification.” International Journal of Remote Sensing 26 (8): 1575–1588. doi:10.1080/01431160512331326765.
  • Erturk, A., D. Cesmeci, M. K. Gullu, D. Gercek, and S. Erturk. 2014. “Endmember Extraction Guided by Anomalies and Homogeneous Regions for Hyperspectral Images.” IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing 7 (8): 3630–3639. doi:10.1109/jstars.2014.2330364.
  • Fauvel, M., J. A. Benediktsson, J. Chanussot, and J. R. Sveinsson. 2008. “Spectral and Spatial Classification of Hyperspectral Data Using Svms and Morphological Profiles.” IEEE Transactions on Geoscience and Remote Sensing 46 (11): 3804–3814. doi:10.1109/TGRS.2008.922034.
  • Fauvel, M., J. Chanussot, and J. A. Benediktsson. 2012. “A Spatial–Spectral Kernel-Based Approach for the Classification of Remote-Sensing Images.” Pattern Recognition 45 (1): 381–392. doi:10.1016/j.patcog.2011.03.035.
  • Fauvel, M., Y. Tarabalka, J. A. Benediktsson, J. Chanussot, and J. C.Tilton. 2013. “Advances in Spectral-Spatial Classification of Hyperspectral Images.” Proceedings of the IEEE 101 (3): 652–675. doi:10.1109/JPROC.2012.2197589.
  • Foody, G. M. 2002. “Status of Land Cover Classification Accuracy Assessment.” Remote Sensing of Environment 80 (1): 185–201. doi:10.1016/S0034-4257(01)00295-4.
  • Foody, G. M. 2008. “Harshness in Image Classification Accuracy Assessment.” International Journal of Remote Sensing 29 (11): 3137–3158. doi:10.1080/01431160701442120.
  • Foody, G. M., A. Mathur, C. Sanchez-Hernandez, and D. S. Boyd. 2006. “Training Set Size Requirements for the Classification of a Specific Class.” Remote Sensing of Environment 104 (1): 1–14. doi:10.1016/j.rse.2006.03.004.
  • Franke, J., D. A. Roberts, K. Halligan, and G. Menz. 2009. “Hierarchical Multiple Endmember Spectral Mixture Analysis (MESMA) of Hyperspectral Imagery for Urban Environments.” Remote Sensing of Environment 113 (8): 1712–1723. doi:10.1016/j.rse.2009.03.018.
  • Fu, K. S., and J. K. Mui. 1981. “A Survey on Image Segmentation.” Pattern Recognition 13 (1): 3–16. doi:10.1016/0031-3203(81)90028-5.
  • García‐Haro, F. J., S. Sommer, and T. Kemper. 2005. “A New Tool for Variable Multiple Endmember Spectral Mixture Analysis (VMESMA).” International Journal of Remote Sensing 26 (10): 2135–2162. doi:10.1080/01431160512331337817.
  • Ge, Y., H. X. Bai, F. Cao, S. P. Li, X. F. Feng, and D. Y. Li. 2009. “Rough Set-Derived Measures in Image Classification Accuracy Assessment.” International Journal of Remote Sensing 30 (20): 5323–5344. doi:10.1080/01431160903131026.
  • Geman, S., and D. Geman. 1984. “Stochastic Relaxation, Gibbs Distributions, and the Bayesian Restoration of Images.” IEEE Transactions on Pattern Analysis and Machine Intelligence PAMI-6 (6): 721–741. doi:10.1109/TPAMI.1984.4767596.
  • Ghamisi, P., J. A.Benediktsson, G. Cavallaro, and A. Plaza. 2014. “Automatic Framework for Spectral–Spatial Classification Based on Supervised Feature Extraction and Morphological Attribute Profiles.” IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing 7 (6): 2147–2160. doi:10.1109/JSTARS.2014.2298876.
  • Ghamisi, P., J. A.Benediktsson, and J. R.Sveinsson. 2014. “Automatic Spectral–Spatial Classification Framework Based on Attribute Profiles and Supervised Feature Extraction.” IEEE Transactions on Geoscience and Remote Sensing 52 (9): 5771–5782. doi:10.1109/TGRS.2013.2292544.
  • Ghamisi, P., M. D. Mura, and J. A.Benediktsson. 2015. “A Survey on Spectral–Spatial Classification Techniques Based on Attribute Profiles.” IEEE Transactions on Geoscience and Remote Sensing 53 (5): 2335–2353. doi:10.1109/TGRS.2014.2358934.
  • Goenaga, M. A., M. C. Torres-Madronero, M. Velez-Reyes, S. J. Van Bloem, and J. D. Chinea. 2013. “Unmixing Analysis of a Time Series of Hyperion Images Over the Guánica Dry Forest in Puerto Rico.” IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing 6 (2): 329–338. doi:10.1109/jstars.2012.2225096.
  • Gonzalez, R. C., R. E. Woods, and S. L. Eddins. 2004. Digital Image Processing Using MATLAB. Lexington: Person Prentice Hall.
  • Greig, D. M., B. T. Porteous, and A. H. Seheult. 1989. “Exact Maximum a Posteriori Estimation for Binary Images.” Journal of the Royal Statistical Society Series B (Methodological) 51 (2): 271–279.
  • Guo, X., X. Huang, and L. Zhang. 2014. “Three-Dimensional Wavelet Texture Feature Extraction and Classification for Multi/Hyperspectral Imagery.” IEEE Geoscience and Remote Sensing Letters 11 (12): 2183–2187. doi:10.1109/LGRS.2014.2323963.
  • Haralick, R. M., K. Shanmugam, and I. Dinstein. 1973. “Textural Features for Image Classification.” IEEE Transactions on Systems, Man and Cybernetics3 (6): 610–621. doi:10.1109/TSMC.1973.4309314.
  • Harsanyi, J. C., and C.-I.Chang. 1994. “Hyperspectral Image Classification and Dimensionality Reduction: An Orthogonal Subspace Projection Approach.” IEEE Transactions on Geoscience and Remote Sensing 32 (4): 779–785. doi:10.1109/36.298007.
  • Hay, G. J., T. Blaschke, D. J.Marceau, and A. Bouchard. 2003. “A Comparison of Three Image-Object Methods for the Multiscale Analysis of Landscape Structure.” ISPRS Journal of Photogrammetry and Remote Sensing 57 (5–6): 327–345. doi:10.1016/S0924-2716(02)00162-4.
  • Hay, G. J., and G. Castilla. 2006. “Object-Based Image Analysis: Strengths, Weaknesses, Opportunities and Threats (SWOT).” International Archives of Photogrammetry, Remote Sensing and Spatial Information Sciences 36 (4): 1–3.
  • Hay, G. J., and G. Castilla. 2008. “Geographic Object-Based Image Analysis (GEOBIA): A New Name for A New Discipline.” In Object-Based Image Analysis, edited by T. Blaschke, S. Lang, and G. Hay, 75–89. Berlin: Springer-Verlag.
  • Hay, G. J., G. Castilla, M. A.Wulder, and J. R.Ruiz. 2005. “An Automated Object-Based Approach for the Multiscale Image Segmentation of Forest Scenes.” International Journal of Applied Earth Observation and Geoinformation 7 (4): 339–359. doi:10.1016/j.jag.2005.06.005.
  • Hay, G. J., D. J. Marceau, P. Dubé, and A. Bouchard. 2001. “A Multiscale Framework for Landscape Analysis: Object-Specific Analysis and Upscaling.” Landscape Ecology 16 (6): 471–490. doi:10.1023/A:1013101931793.
  • Heinz, D. C., and C.-I.-C.Chang. 2001. “Fully Constrained Least Squares Linear Spectral Mixture Analysis Method for Material Quantification in Hyperspectral Imagery.” IEEE Transactions on Geoscience and Remote Sensing 39 (3): 529–545. doi:10.1109/36.911111.
  • Hernando, A., D. Tiede, F. Albrecht, and S. Lang. 2012. “Spatial and Thematic Assessment of Object-Based Forest Stand Delineation Using an OFA-Matrix.” International Journal of Applied Earth Observation and Geoinformation 19: 214–225. doi:10.1016/j.jag.2012.05.007.
  • Herzfeld, U. C., and C. A.Higginson. 1996. “Automated Geostatistical Seafloor Classification - Principles, Parameters, Feature Vectors, and Discrimination Criteria.” Computers & Geosciences 22 (1): 35–52. doi:10.1016/0098-3004(96)89522-7.
  • Hoberg, T., F. Rottensteiner, R. Q. Feitosa, and C. Heipke. 2015. “Conditional Random Fields for Multitemporal and Multiscale Classification of Optical Satellite Imagery.” IEEE Transactions on Geoscience and Remote Sensing 53 (2): 659–673. doi:10.1109/TGRS.2014.2326886.
  • Huang, X., X. Liu, and L. Zhang. 2014a. “A Multichannel Gray Level Co-Occurrence Matrix for Multi/Hyperspectral Image Texture Representation.” Remote Sensing 6 (9): 8424–8445. doi:10.3390/rs6098424.
  • Huang, X., Q. Lu, L. Zhang, and A. Plaza. 2014b. “New Postprocessing Methods for Remote Sensing Image Classification: A Systematic Study.” IEEE Transactions on Geoscience and Remote Sensing 52 (11): 7140–7159. doi:10.1109/TGRS.2014.2308192.
  • Huang, X., and L. Zhang. 2012. “A Multiscale Urban Complexity Index Based on 3D Wavelet Transform for Spectral–Spatial Feature Extraction and Classification: An Evaluation on the 8-Channel Worldview-2 Imagery.” International Journal of Remote Sensing 33 (8): 2641–2656. doi:10.1080/01431161.2011.614287.
  • Huang, X., L. Zhang, and P. Li. 2008. “A Multiscale Feature Fusion Approach for Classification of Very High Resolution Satellite Imagery Based on Wavelet Transform.” International Journal of Remote Sensing 29 (20): 5923–5941. doi:10.1080/01431160802139922.
  • Huang, X., L. Zhang, and P. Li. 2007. “Classification and Extraction of Spatial Features in Urban Areas Using High-Resolution Multispectral Imagery.” IEEE Geoscience and Remote Sensing Letters 4 (2): 260–264. doi:10.1109/LGRS.2006.890540.
  • Ilea, D. E., and P. F.Whelan. 2011. “Image Segmentation Based on the Integration of Colour–Texture Descriptors—A Review.” Pattern Recognition 44 (10–11): 2479–2501. doi:10.1016/j.patcog.2011.03.005.
  • Iordache, M.-D., J. M. Bioucas-Dias, and A. Plaza. 2011. “Sparse Unmixing of Hyperspectral Data.” IEEE Transactions on Geoscience and Remote Sensing 49 (6): 2014–2039. doi:10.1109/tgrs.2010.2098413.
  • Iordache, M.-D., J. M. Bioucas-Dias, and A. Plaza. 2012. “Total Variation Spatial Regularization for Sparse Hyperspectral Unmixing.” IEEE Transactions on Geoscience and Remote Sensing 50 (11): 4484–4502. doi:10.1109/tgrs.2012.2191590.
  • Jackson, Q., and D. A.Landgrebe. 2002. “Adaptive Bayesian Contextual Classification Based on Markov Random Fields.” IEEE Transactions on Geoscience and Remote Sensing 40 (11): 2454–2463. doi:10.1109/TGRS.2002.805087.
  • Jain, A., and G. Healey. 1998. “A Multiscale Representation Including Opponent Color Features for Texture Recognition.” IEEE Transactions on Image Processing 7 (1): 124–128. doi:10.1109/83.650858.
  • Jensen, J. R. 2000. Remote Sensing of the Environment: An Earth Resource Perspective. Upper Saddle River, NJ: Prentice Hall.
  • Jensen, J. R. 1996. Introductory Digital Image Processing: A Remote Sensing Perspective. 2nd ed. Upper Saddle River, NJ: Prentice Hall.
  • Jensen, J. R. 2009. Remote Sensing of the Environment: An Earth Resource Perspective 2/E. Pearson Education India.
  • Jensen, J. R., and K. Lulla. 1987. Introductory Digital Image Processing: A Remote Sensing Perspective.
  • Jia, S., and Y. T. Qian. 2007. “Spectral and Spatial Complexity-Based Hyperspectral Unmixing.” IEEE Transactions on Geoscience and Remote Sensing 45 (12): 3867–3879. doi:10.1109/tgrs.2007.898443.
  • Jia, S., and Y. T. Qian. 2009. “Constrained Nonnegative Matrix Factorization for Hyperspectral Unmixing.” IEEE Transactions on Geoscience and Remote Sensing 47 (1): 161–173. doi:10.1109/tgrs.2008.2002882.
  • Jia, X., and J. A. Richards. 2008. “Managing the Spectral-Spatial Mix in Context Classification Using Markov Random Fields.” IEEE Geoscience and Remote Sensing Letters 5 (2): 311–314. doi:10.1109/LGRS.2008.916076.
  • Johnson, B., R. Tateishi, and T. Kobayashi. 2012. “Remote Sensing of Fractional Green Vegetation Cover Using Spatially-Interpolated Endmembers.” Remote Sensing 4 (12): 2619–2634. doi:10.3390/rs4092619.
  • Keshava, N., and J. F. Mustard. 2002. “Spectral Unmixing.” IEEE Signal Processing Magazine 19 (1): 44–57. doi:10.1109/79.974727.
  • Kim, M., T. A.Warner, M. Madden, and D. S.Atkinson. 2011. “Multi-Scale GEOBIA with Very High Spatial Resolution Digital Aerial Imagery: Scale, Texture and Image Objects.” International Journal of Remote Sensing 32 (10): 2825–2850. doi:10.1080/01431161003745608.
  • Kindermann, R., and J. L. Snell. 1980. Markov Random Fields and Their Applications. Vol. 1. Providence, RI: American Mathematical Society.
  • King, R. L., and N. H. Younan. 2006. “Pixel Unmixing via Information of Neighboring Pixels.” GIScience & Remote Sensing 43 (4): 310–322. doi:10.2747/1548-1603.43.4.310.
  • Kolmogorov, V. 2006. “Convergent Tree-Reweighted Message Passing for Energy Minimization.” IEEE Transactions on Pattern Analysis and Machine Intelligence 28 (10): 1568–1583. doi:10.1109/TPAMI.2006.200.
  • Kolmogorov, V., and R. Zabih. 2004. “What Energy Functions Can Be Minimized via Graph Cuts?” IEEE Transactions on Pattern Analysis and Machine Intelligence 26 (2): 147–159. doi:10.1109/TPAMI.2004.1262177.
  • Kosov, S., F. Rottensteiner, and C. Heipke. 2013. “Sequential Gaussian Mixture Models for two-level Conditional Random Fields.” In Pattern Recognition, 153–163. Berlin: Springer.
  • Kumar, S., and M. Hebert. 2003. “Discriminative Random Fields: A Discriminative Framework for Contextual Interaction in Classification.” Paper presented at the Computer Vision, 2003. Proceedings of Ninth IEEE International Conference on Computer Vision, Nice.
  • Kyriakidis, P. C., and J. L.Dungan. 2001. “A Geostatistical Approach for Mapping Thematic Classification Accuracy and Evaluating the Impact of Inaccurate Spatial Data on Ecological Model Predictions.” Environmental and Ecological Statistics 8 (4): 311–330. doi:10.1023/A:1012778302005.
  • Lafferty, J., A. McCallum, and F. C. N. Pereira. 2001.” Conditional Random Fields: Probabilistic Models for Segmenting and Labeling Sequence Data.“ In: Proceedings of the International Conference on Machine Learning (ICML’01), Williamstown, MA, 282–289.
  • Lang, S. 2008. “Object-Based Image Analysis for Remote Sensing Applications: Modeling Reality–Dealing with Complexity.” In Object-Based Image Analysis, edited by T. Blaschke, S. Lang, and G. J. Hay, 3–27. Berlin: Springer.
  • Li, H. L., and L. P. Zhang. 2011a. “A Hybrid Automatic Endmember Extraction Algorithm Based on a Local Window.” IEEE Transactions on Geoscience and Remote Sensing 49 (11): 4223–4238. doi:10.1109/tgrs.2011.2162098.
  • Li, M., S. Y. Zang, C. S. Wu, and Y. B. Deng. 2015. “Segmentation-Based and Rule-Based Spectral Mixture Analysis for Estimating Urban Imperviousness.” Advances in Space Research 55 (5): 1307–1315. doi:10.1016/j.asr.2014.12.015.
  • Li, P., T. Cheng, and J. Guo. 2009. “Multivariate Image Texture by Multivariate Variogram for Multispectral Image Classification.” Photogrammetric Engineering & Remote Sensing 75 (2): 147–157. doi:10.14358/PERS.75.2.147.
  • Li, S. Z. 2009. Markov Random Field Modeling in Image Analysis. 3rd ed. London: Springer.
  • Li, W. D., and C. R. Zhang. 2011b. “A Markov Chain Geostatistical Framework for Land-Cover Classification With Uncertainty Assessment Based on Expert-Interpreted Pixels From Remotely Sensed Imagery.” IEEE Transactions on Geoscience and Remote Sensing 49 (8): 2983–2992. doi:10.1109/TGRS.2011.2121916.
  • Liao, W., R. Bellens, A. Pizurica, W. Philips, and Y. Pi. 2012. “Classification of Hyperspectral Data over Urban Areas Using Directional Morphological Profiles and Semi-Supervised Feature Extraction.” IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing 5 (4): 1177–1190. doi:10.1109/JSTARS.2012.2190045.
  • Liu, J. M., J. S. Zhang, Y. L. Gao, C. X. Zhang, and Z. H. Li. 2012. “Enhancing Spectral Unmixing by Local Neighborhood Weights.” IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing 5 (5): 1545–1552. doi:10.1109/jstars.2012.2199282.
  • Liu, J., Z. Wu, Z. Wei, L. Xiao, and L. Sun, 2013. “Spatial-Spectral Kernel Sparse Representation for Hyperspectral Image Classification.“ IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing 6 (6): 2462–2471. doi:10.1109/JSTARS.2013.2252150
  • Liu, X. S., W. Xia, B. Wang, and L. M. Zhang. 2011. “An Approach Based on Constrained Nonnegative Matrix Factorization to Unmix Hyperspectral Data.” IEEE Transactions on Geoscience and Remote Sensing 49 (2): 757–772. doi:10.1109/tgrs.2010.2068053.
  • Liu, Y., Q. Guo, and M. Kelly. 2008. “A Framework of Region-Based Spatial Relations for Non-Overlapping Features and Its Application in Object Based Image Analysis.” ISPRS Journal of Photogrammetry and Remote Sensing 63 (4): 461–475. doi:10.1016/j.isprsjprs.2008.01.007.
  • Lizarazo, I. 2014. “Accuracy Assessment of Object-Based Image Classification: Another STEP.” International Journal of Remote Sensing 35 (16): 6135–6156. doi:10.1080/01431161.2014.943328.
  • Lloyd, C. D., S. Berberoglu, P. J. Curran, and P. M. Atkinson. 2004. “A Comparison of Texture Measures for the Per-Field Classification of Mediterranean Land Cover.” International Journal of Remote Sensing 25 (19): 3943–3965. doi:10.1080/0143116042000192321.
  • Lopez, S., J. F. Moure, A. Plaza, G. M. Callico, J. F. Lopez, and R. Sarmiento. 2013. “A New Preprocessing Technique for Fast Hyperspectral Endmember Extraction.” IEEE Geoscience and Remote Sensing Letters 10 (5): 1070–1074. doi:10.1109/lgrs.2012.2229689.
  • Lu, D., and Q. Weng. 2007. “A Survey of Image Classification Methods and Techniques for Improving Classification Performance.” International Journal of Remote Sensing 28 (5): 823–870. doi:10.1080/01431160600746456.
  • Ma, B. D., L. X. Wu, X. X. Zhang, X. C. Li, Y. Liu, and S. L. Wang. 2014. “Locally Adaptive Unmixing Method for Lake-Water Area Extraction Based on MODIS 250 M Bands.” International Journal of Applied Earth Observation and Geoinformation 33: 109–118. doi:10.1016/j.jag.2014.05.002.
  • MacLean, M. G., M. J. Campbell, D. S. Maynard, M. J. Ducey, and R. G. Congalton. 2013. “Requirements for Labelling Forest Polygons in an Object-Based Image Analysis Classification.” International Journal of Remote Sensing 34 (7): 2531–2547. doi:10.1080/01431161.2012.747017.
  • Magnussen, S., and S. de Bruin. 2003. “Updating Cover Type Maps Using Sequential Indicator Simulation.” Remote Sensing of Environment 87 (2–3): 161–170. doi:10.1016/S0034-4257(03)00138-X.
  • Maillard, P. 2003. “Comparing Texture Analysis Methods through Classification.” Photogrammetric Engineering & Remote Sensing 69 (4): 357–367. doi:10.14358/PERS.69.4.357.
  • Mallinis, G., N. Koutsias, M. Tsakiri-Strati, and M. Karteris. 2008. “Object-Based Classification Using Quickbird Imagery for Delineating Forest Vegetation Polygons in a Mediterranean Test Site.” ISPRS Journal of Photogrammetry and Remote Sensing 63 (2): 237–250. doi:10.1016/j.isprsjprs.2007.08.007.
  • Martin, G., and A. Plaza. 2011. “Region-Based Spatial Preprocessing for Endmember Extraction and Spectral Unmixing.” IEEE Geoscience and Remote Sensing Letters 8 (4): 745–749. doi:10.1109/lgrs.2011.2107877.
  • Martin, G., and A. Plaza. 2012. “Spatial-Spectral Preprocessing Prior to Endmember Identification and Unmixing of Remotely Sensed Hyperspectral Data.” IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing 5 (2): 380–395. doi:10.1109/jstars.2012.2192472.
  • Maselli, F. 2001. “Definition of Spatially Variable Spectral Endmembers by Locally Calibrated Multivariate Regression Analyses.” Remote Sensing of Environment 75 (1): 29–38. doi:10.1016/s0034-4257(00)00153-x.
  • McGarigal, K., S. A. Cushman, and E. Ene. 2012. “FRAGSTATS v4: Spatial Pattern Analysis Program for Categorical and Continuous Maps.” Computer Software Program Produced by the Authors at the University of Massachusetts, Amherst. http://www.umass.edu/landeco/research/fragstats/fragstats.html.”
  • Mei, S. H., Q. Du, and M. Y. He. 2015. “Equivalent-Sparse Unmixing Through Spatial and Spectral Constrained Endmember Selection From an Image-Derived Spectral Library.” IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing 8 (6): 2665–2675. doi:10.1109/jstars.2015.2403254.
  • Mei, S. H., M. Y. He, Z. Y. Wang, and D. G. Feng. 2010. “Spatial Purity Based Endmember Extraction for Spectral Mixture Analysis.” IEEE Transactions on Geoscience and Remote Sensing 48 (9): 3434–3445. doi:10.1109/tgrs.2010.2046671.
  • Mei, S. H., M. Y. He, Y. F. Zhang, Z. Y. Wang, and D. G. Feng. 2011. “Improving Spatial–Spectral Endmember Extraction in the Presence of Anomalous Ground Objects.” IEEE Transactions on Geoscience and Remote Sensing 49 (11): 4210–4222. doi:10.1109/tgrs.2011.2163160.
  • Melgani, F., and S. B.Serpico. 2003. “A Markov Random Field Approach to Spatio-Temporal Contextual Image Classification.” IEEE Transactions on Geoscience and Remote Sensing 41 (11): 2478–2487. doi:10.1109/TGRS.2003.817269.
  • Möller, M., J. Birger, A. Gidudu, and C. Gläßer. 2013. “A Framework for the Geometric Accuracy Assessment of Classified Objects.” International Journal of Remote Sensing 34 (24): 8685–8698. doi:10.1080/01431161.2013.845319.
  • Moser, G., and S. B.Serpico. 2013. “Combining Support Vector Machines and Markov Random Fields in an Integrated Framework for Contextual Image Classification.” IEEE Transactions on Geoscience and Remote Sensing51 (5): 2734–2752. doi:10.1109/TGRS.2012.2211882.
  • Moser, G., S. B.Serpico, and J. A.Benediktsson. 2013. “Land-Cover Mapping by Markov Modeling of Spatial–Contextual Information in Very-High-Resolution Remote Sensing Images.” Proceedings of the IEEE 101 (3): 631–651. doi:10.1109/JPROC.2012.2211551.
  • Moser, G., and S. B. Serpico. 2008. “Classification of High-Resolution Images Based on MRF Fusion and Multiscale Segmentation.” Paper presented at the Geoscience and Remote Sensing Symposium, 2008. IGARSS 2008. IEEE International, Boston, MA.
  • Moser, G., and S. B. Serpico. 2011. “Contextual High-Resolution Image Classification by Markovian Data Fusion, Adaptive Texture Extraction, and Multiscale Segmentation.” Paper presented at the 2011 IEEE International on Geoscience and Remote Sensing Symposium (IGARSS), 2011 IEEE International, Vancouver, BC.
  • Mountrakis, G., and B. Xi. 2013. “Assessing Reference Dataset Representativeness through Confidence Metrics Based on Information Density.” ISPRS Journal of Photogrammetry and Remote Sensing 78: 129–147. doi:10.1016/j.isprsjprs.2013.01.011.
  • Mountrakis, G., J. Im, and C. Ogole. 2011. “Support Vector Machines in Remote Sensing: A Review.” ISPRS Journal of Photogrammetry and Remote Sensing 66 (3): 247–259. doi:10.1016/j.isprsjprs.2010.11.001.
  • Myint, S. W.. 2003. “Fractal Approaches in Texture Analysis and Classification of Remotely Sensed Data: Comparisons with Spatial Autocorrelation Techniques and Simple Descriptive Statistics.” International Journal of Remote Sensing 24 (9): 1925–1947. doi:10.1080/01431160210155992.
  • Myint, S. W., and N. Lam. 2005. “Examining Lacunarity Approaches in Comparison with Fractal and Spatial Autocorrelation Techniques for Urban Mapping.” Photogrammetric Engineering & Remote Sensing 71 (8): 927–937. doi:10.14358/PERS.71.8.927.
  • Myint, S. W., N. S.-N. Lam, and J. M.Tyler. 2004. “Wavelets for Urban Spatial Feature Discrimination: Comparisons with Fractal, Spatial Autocorrelation, and Spatial Co-Occurrence Approaches.” Photogrammetric Engineering & Remote Sensing 70 (7): 803–812. doi:10.14358/PERS.70.7.803.
  • Nascimento, J. M. P., and J. M. B. Dias. 2005. “Vertex Component Analysis: A Fast Algorithm to Unmix Hyperspectral Data.” IEEE Transactions on Geoscience and Remote Sensing 43 (4): 898–910. doi:10.1109/tgrs.2005.844293.
  • Neubert, M., H. Herold, and G. Meinel. 2008. “Assessing Image Segmentation Quality–Concepts, Methods and Application.” In Object-Based Image Analysis, edited by T. Blaschke, S. Lang, and G. J. Hay, 769–784. Berlin: Springer-Verlag.
  • Neville, R. A., K. Staenz, T. Szeredi, J. Lefebvre, and P. Hauff. 1999. “Automatic Endmember Extraction from Hyperspectral Data for Mineral Exploration.” Paper presented at the Fourth international airborne remote sensing conference and exhibition /21st Canadian symposium on remote sensing, Ottawa, Ontario, Canada.
  • Ojala, T., and M. Pietikäinen. 1999. “Unsupervised Texture Segmentation Using Feature Distributions.” Pattern Recognition 32 (3): 477–486. doi:10.1016/S0031-3203(98)00038-7.
  • Pal, N. R., and S. K.Pal. 1993. “A Review on Image Segmentation Techniques.” Pattern Recognition 26 (9): 1277–1294. doi:10.1016/0031-3203(93)90135-J.
  • Pant, T., D. Singh, and T. Srivastava. 2010. “Advanced Fractal Approach for Unsupervised Classification of SAR Images.” Advances in Space Research 45 (11): 1338–1349. doi:10.1016/j.asr.2010.01.008.
  • Park, N.-W., K.-H.Chi, and B.-D.Kwon. 2003. “Geostatistical Integration of Spectral and Spatial Information for Land-Cover Mapping Using Remote Sensing Data.” Geosciences Journal 7 (4): 335–341. doi:10.1007/BF02919565.
  • Parrinello, T., and R. A. Vaughan. 2002. “Multifractal Analysis and Feature Extraction in Satellite Imagery.” International Journal of Remote Sensing 23 (9): 1799–1825. doi:10.1080/01431160110075820.
  • Parrinello, T., and R. A. Vaughan. 2006. “On Comparing Multifractal and Classical Features in Minimum Distance Classification of AVHRR Imagery.” International Journal of Remote Sensing 27 (18): 3943–3959. doi:10.1080/01431160600685241.
  • Persello, C., and L. Bruzzone. 2010. “A Novel Protocol for Accuracy Assessment in Classification of Very High Resolution Images.” IEEE Transactions on Geoscience and Remote Sensing 48 (3): 1232–1244. doi:10.1109/TGRS.2009.2029570.
  • Pesaresi, M., and J. A.Benediktsson. 2001. “A New Approach for the Morphological Segmentation of High-Resolution Satellite Imagery.” IEEE Transactions on Geoscience and Remote Sensing 39 (2): 309–320. doi:10.1109/36.905239.
  • Plaza, A., P. Martinez, R. Perez, and J. Plaza. 2002. “Spatial/Spectral Endmember Extraction by Multidimensional Morphological Operations.” IEEE Transactions on Geoscience and Remote Sensing 40 (9): 2025–2041. doi:10.1109/tgrs.2002.802494.
  • Plaza, A., J. A.Benediktsson, J. W.Boardman, J. Brazile, L. Bruzzone, G. Camps-Valls, J. Chanussot, et al. 2009. “Recent Advances in Techniques for Hyperspectral Image Processing.” Remote Sensing of Environment 113: SS110-SS122. doi:10.1016/j.rse.2007.07.028.
  • Pontius, R. G.Jr, and M. Millones. 2011. “Death to Kappa: Birth of Quantity Disagreement and Allocation Disagreement for Accuracy Assessment.” International Journal of Remote Sensing 32 (15): 4407–4429. doi:10.1080/01431161.2011.552923.
  • Potere, D., A. Schneider, S. Angel, and D. L. Civco. 2009. “Mapping Urban Areas on a Global Scale: Which of the Eight Maps Now Available Is More Accurate?” International Journal of Remote Sensing 30 (24): 6531–6558. doi:10.1080/01431160903121134.
  • Power, C., A. Simms, and R. White. 2001. “Hierarchical Fuzzy Pattern Matching for the Regional Comparison of Land Use Maps.” International Journal of Geographical Information Science 15 (1): 77–100. doi:10.1080/136588100750058715.
  • Radoux, J., and P. Bogaert. 2014. “Accounting for the Area of Polygon Sampling Units for the Prediction of Primary Accuracy Assessment Indices.” Remote Sensing of Environment 142: 9–19. doi:10.1016/j.rse.2013.10.030.
  • Radoux, J., P. Bogaert, D. Fasbender, and P. Defourny. 2011. “Thematic Accuracy Assessment of Geographic Object-Based Image Classification.” International Journal of Geographical Information Science 25 (6): 895–911. doi:10.1080/13658816.2010.498378.
  • Rand, R. S., and D. M. Keenan. 2001. “A Spectral Mixture Process Conditioned by Gibbs-Based Partitioning.” IEEE Transactions on Geoscience and Remote Sensing 39 (7): 1421–1434. doi:10.1109/36.934074.
  • Roberts, D. A., M. Gardner, R. Church, S. Ustin, G. Scheer, and R. O. Green. 1998. “Mapping Chaparral in the Santa Monica Mountains Using Multiple Endmember Spectral Mixture Models.” Remote Sensing of Environment 65 (3): 267–279. doi:10.1016/S0034-4257(98)00037-6.
  • Rodriguez-Galiano, V. F., M. Chica-Olmo, F. Abarca-Hernandez, P. M. Atkinson, and C. Jeganathan. 2012. “Random Forest Classification of Mediterranean Land Cover Using Multi-Seasonal Imagery and Multi-Seasonal Texture.” Remote Sensing of Environment 121: 93–107. doi:10.1016/j.rse.2011.12.003.
  • Roessner, S., K. Segl, U. Heiden, and H. Kaufmann. 2001. “Automated Differentiation of Urban Surfaces Based on Airborne Hyperspectral Imagery.” IEEE Transactions on Geoscience and Remote Sensing 39 (7): 1525–1532. doi:10.1109/36.934082.
  • Rogge, D., M. Bachmann, B. Rivard, and J. L. Feng. 2012. “Spatial Sub-Sampling Using Local Endmembers for Adapting OSP and SSEE for Large-Scale Hyperspectral Surveys.” IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing 5 (1): 183–195. doi:10.1109/jstars.2011.2168513.
  • Rogge, D. M., B. Rivard, J. Zhang, A. Sanchez, J. Harris, and J. Feng. 2007. “Integration of Spatial-Spectral Information for the Improved Extraction of Endmembers.” Remote Sensing of Environment 110 (3): 287–303. doi:10.1016/j.rse.2007.02.019.
  • Rubinstein, R., A. M.Bruckstein, and M. Elad. 2010. “Dictionaries for Sparse Representation Modeling.” Proceedings of the IEEE 98 (6): 1045–1057. doi:10.1109/JPROC.2010.2040551.
  • Schiewe, J. 2002. “Segmentation of High-Resolution Remotely Sensed Data-Concepts, Applications and Problems.” International Archives of Photogrammetry Remote Sensing and Spatial Information Sciences 34 (4): 380–385.
  • Schindler, K. 2012. “An Overview and Comparison of Smooth Labeling Methods for Land-Cover Classification.” IEEE Transactions on Geoscience and Remote Sensing50 (11): 4534–4545. doi:10.1109/TGRS.2012.2192741.
  • Schölkopf, B., and A. J. Smola. 2002. Learning with Kernels: Support Vector Machines, Regularization, Optimization, and Beyond. Cambridge, MA: MIT Press.
  • Schöpfer, E., and S. Lang. 2006. “Object Fate Analysis–A Virtual Overlay Method for the Categorisation of Object Transition and Object-Based Accuracy Assessment.” International Archives of Photogrammetry, Remote Sensing and Spatial Information Sciences 36 (4): C42.
  • Settle, J. J., and N. A. Drake. 1993. “Linear Mixing and the Estimation of Ground Cover Proportions.” International Journal of Remote Sensing 14 (6): 1159–1177. doi:10.1080/01431169308904402.
  • Shen, L., and S. Jia. 2011. “Three-Dimensional Gabor Wavelets for Pixel-Based Hyperspectral Imagery Classification.” IEEE Transactions on Geoscience and Remote Sensing 49 (12): 5039–5046. doi:10.1109/TGRS.2011.2157166.
  • Shen, L., Z. Zhu, S. Jia, J. Zhu, and Y. Sun. 2013. “Discriminative Gabor Feature Selection for Hyperspectral Image Classification.” IEEE Geoscience and Remote Sensing Letters 10 (1): 29–33. doi:10.1109/LGRS.2012.2191761.
  • Shi, C., and L. Wang. 2014. “Incorporating Spatial Information in Spectral Unmixing: A Review.” Remote Sensing of Environment 149: 70–87. doi:10.1016/j.rse.2014.03.034.
  • Shi, M., and G. Healey. 2003. “Hyperspectral Texture Recognition Using a Multiscale Opponent Representation.” IEEE Transactions on Geoscience and Remote Sensing 41 (5): 1090–1095. doi:10.1109/TGRS.2003.811076.
  • Shimabukuro, Y. E., and J. A. Smith. 1991. “The Least-Squares Mixing Models to Generate Fraction Images Derived from Remote-Sensing Multispectral Data.” IEEE Transactions on Geoscience and Remote Sensing 29 (1): 16–20. doi:10.1109/36.103288.
  • Shoshany, M., and T. Svoray. 2002. “Multidate Adaptive Unmixing and Its Application to Analysis of Ecosystem Transitions along a Climatic Gradient.” Remote Sensing of Environment 82 (1): 5–20. doi:10.1016/s0034-4257(01)00346-7.
  • Silvetti, A. F., and C. A.Delrieux. 2013. “Quadratic Self-Correlation: An Improved Method for Computing Local Fractal Dimension in Remote Sensing Imagery.” Computers & Geosciences 60: 142–155. doi:10.1016/j.cageo.2013.06.008.
  • Soille, P., and M. Pesaresi. 2002. “Advances in Mathematical Morphology Applied to Geoscience and Remote Sensing.” IEEE Transactions on Geoscience and Remote Sensing 40 (9): 2042–2055. doi:10.1109/TGRS.2002.804618.
  • Solberg, A. H. S., T. Taxt, and A. K.Jain. 1996. “A Markov Random Field Model for Classification of Multisource Satellite Imagery.” IEEE Transactions on Geoscience and Remote Sensing 34 (1): 100–113. doi:10.1109/36.481897.
  • Soltani-Farani, A., and H. R. Rabiee. 2015. “When Pixels Team up: Spatially Weighted Sparse Coding for Hyperspectral Image Classification.” IEEE Geoscience and Remote Sensing Letters 12 (1): 107–111. doi:10.1109/lgrs.2014.2328319.
  • Somers, B., G. P. Asner, L. Tits, and P. Coppin. 2011. “Endmember Variability in Spectral Mixture Analysis: A Review.” Remote Sensing of Environment 115 (7): 1603–1616. doi:10.1016/j.rse.2011.03.003.
  • Song, B., J. Li, M. Dalla Mura, P. Li, A. Plaza, J. M.Bioucas-Dias, J .Atli Benediktsson, and J. Chanussot. 2014. “Remotely Sensed Image Classification Using Sparse Representations of Morphological Attribute Profiles.” IEEE Transactions on Geoscience and Remote Sensing 52 (8): 5122–5136. doi:10.1109/TGRS.2013.2286953.
  • Song, X., X. Jiang, and X. Rui. 2010. “Spectral Unmixing Using Linear Unmixing under Spatial Autocorrelation Constraints.” Paper presented at the 2010 IEEE International on Geoscience and Remote Sensing Symposium (IGARSS), July 25–30, Honolulu, HI.
  • Stehman, S. V.2009. “Sampling Designs for Accuracy Assessment of Land Cover.” International Journal of Remote Sensing 30 (20): 5243–5272. doi:10.1080/01431160903131000.
  • Stehman, S. V., and R. L.Czaplewski. 1998. “Design and Analysis for Thematic Map Accuracy Assessment: Fundamental Principles.” Remote Sensing of Environment 64 (3): 331–344. doi:10.1016/S0034-4257(98)00010-8.
  • Stehman, S. V., and J. D. Wickham. 2011. “Pixels, Blocks of Pixels, and Polygons: Choosing a Spatial Unit for Thematic Accuracy Assessment.” Remote Sensing of Environment 115 (12): 3044–3055. doi:10.1016/j.rse.2011.06.007.
  • Sun, W., G. Xu, P. Gong, and S. Liang. 2006. “Fractal Analysis of Remotely Sensed Images: A Review of Methods and Applications.” International Journal of Remote Sensing 27 (22): 4963–4990. doi:10.1080/01431160600676695.
  • Sun, W. 2006. “Three New Implementations of the Triangular Prism Method for Computing the Fractal Dimension of Remote Sensing Images.” Photogrammetric Engineering & Remote Sensing 72 (4): 373–382. doi:10.14358/PERS.72.4.373.
  • Suykens, J. A. K., and J. Vandewalle. 1999. “Least Squares Support Vector Machine Classifiers.” Neural Processing Letters 9 (3): 293–300. doi:10.1023/A:1018628609742.
  • Szeliski, R., R. Zabih, D. Scharstein, O. Veksler, V. Kolmogorov, A. Agarwala, M. Tappen, and C. Rother. 2008. “A Comparative Study of Energy Minimization Methods for Markov Random Fields with Smoothness-Based Priors.” IEEE Transactions on Pattern Analysis and Machine Intelligence 30 (6): 1068–1080. doi:10.1109/TPAMI.2007.70844.
  • Tarabalka, Y., J. A.Benediktsson, and J. Chanussot. 2009. “Spectral–Spatial Classification of Hyperspectral Imagery Based on Partitional Clustering Techniques.” IEEE Transactions on Geoscience and Remote Sensing 47 (8): 2973–2987. doi:10.1109/TGRS.2009.2016214.
  • Tarabalka, Y., J. A. Benediktsson, J. Chanussot, and J. C. Tilton. 2010a. “Multiple Spectral–Spatial Classification Approach for Hyperspectral Data.” IEEE Transactions on Geoscience and Remote Sensing 48 (11): 4122–4132.
  • Tarabalka, Y., J. Chanussot, and J. A.Benediktsson. 2010. “Segmentation and Classification of Hyperspectral Images Using Watershed Transformation.” Pattern Recognition 43 (7): 2367–2379. doi:10.1016/j.patcog.2010.01.016.
  • Tarabalka, Y., M. Fauvel, J. Chanussot, and J. A.Benediktsson. 2010b. “SVM- and MRF-Based Method for Accurate Classification of Hyperspectral Images.” IEEE Geoscience and Remote Sensing Letters 7 (4): 736–740. doi:10.1109/LGRS.2010.2047711.
  • Tarabalka, Y., J. C.Tilton, J. A.Benediktsson, and J. Chanussot. 2012. “A Marker-Based Approach for the Automated Selection of A Single Segmentation from A Hierarchical Set of Image Segmentations.” IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing 5 (1): 262–272. doi:10.1109/JSTARS.2011.2173466.
  • Taubenbock, H., T. Esch, A. Felbier, A. Roth, and S. Dech. 2011. “Pattern-Based Accuracy Assessment of an Urban Footprint Classification Using Terrasar-X Data.” IEEE Geoscience and Remote Sensing Letters8 (2): 278–282. doi:10.1109/LGRS.2010.2069083.
  • Thompson, D. R., L. Mandrake, M. S. Gilmore, and R. Castano. 2010. “Superpixel Endmember Detection.” IEEE Transactions on Geoscience and Remote Sensing 48 (11): 4023–4033. doi:10.1109/tgrs.2010.2070802.
  • Tiede, D., S. Lang, and C. Hoffmann. 2008. “Domain-Specific Class Modelling for One-Level Representation of Single Trees.” In Object-Based Image Analysis, edited by T. Blaschke, S. Lang, and G. Hay, 133–151. New York: Springer.
  • Tilton, J. C. 1998. “Image Segmentation by Region Growing and Spectral Clustering with Natural Convergence Criterion.” Paper Presented at the International Geoscience and Remote Sensing Symposium, Seattle, WA.
  • Tilton, J. C., Y. Tarabalka, P. M.Montesano, and E. Gofman. 2012. “Best Merge Region-Growing Segmentation with Integrated Nonadjacent Region Object Aggregation.” IEEE Transactions on Geoscience and Remote Sensing 50 (11): 4454–4467. doi:10.1109/TGRS.2012.2190079.
  • Torres-Madronero, M. C., and M. Velez-Reyes. 2014. “Integrating Spatial Information in Unsupervised Unmixing of Hyperspectral Imagery Using Multiscale Representation.” IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing 7 (6): 1985–1993. doi:10.1109/jstars.2014.2319261.
  • Tsai, F., and J.-S.Lai. 2013. “Feature Extraction of Hyperspectral Image Cubes Using Three-Dimensional Gray-Level Cooccurrence.” IEEE Transactions on Geoscience and Remote Sensing 51 (6): 3504–3513. doi:10.1109/TGRS.2012.2223704.
  • Tso, B., and R. C.Olsen. 2005. “A Contextual Classification Scheme Based on MRF Model with Improved Parameter Estimation and Multiscale Fuzzy Line Process.” Remote Sensing of Environment 97 (1): 127–136. doi:10.1016/j.rse.2005.04.021.
  • Tuia, D., G. Camps-Valls, G. Matasci, and M. Kanevski. 2010a. “Learning Relevant Image Features with Multiple-Kernel Classification.” IEEE Transactions on Geoscience and Remote Sensing48 (10): 3780–3791. doi:10.1109/TGRS.2010.2049496.
  • Tuia, D., F. Ratle, A. Pozdnoukhov, and G. Camps-Valls. 2010b. “Multisource Composite Kernels for Urban-Image Classification.” IEEE Geoscience and Remote Sensing Letters 7 (1): 88–92. doi:10.1109/LGRS.2009.2015341.
  • Van Der Kwast, J., T. Van De Voorde, F. Canters, I. Uljee, S. Van Looy, and G. Engelen. 2011. “Inferring Urban Land Use Using the Optimised Spatial Reclassification Kernel.” Environmental Modelling & Software 26 (11): 1279–1288. doi:10.1016/j.envsoft.2011.05.012.
  • Van Der Meer, F. 1994. “Extraction of Mineral Absorption Features from High-Spectralresolution Data Using Non-Parametric Geostatistical Techniques.” International Journal of Remote Sensing 15 (11): 2193–2214. doi:10.1080/01431169408954238.
  • Van Der Meer, F. 1999. “Iterative Spectral Unmixing (ISU).” International Journal of Remote Sensing 20 (17): 3431–3436. doi:10.1080/014311699211462.
  • Van Der Meer, F. 2012. “Remote-Sensing Image Analysis and Geostatistics.” International Journal of Remote Sensing 33 (18): 5644–5676. doi:10.1080/01431161.2012.666363.
  • Vapnik, V. 2013. The Nature of Statistical Learning Theory. New York: Springer Science & Business Media.
  • Wainwright, M. J., T. S.Jaakkola, and A. S.Willsky. 2005. “MAP Estimation via Agreement on Trees: Message-Passing and Linear Programming.” IEEE Transactions on Information Theory 51 (11): 3697–3717. doi:10.1109/TIT.2005.856938.
  • Wallach, H. M. 2004. Conditional Random Fields: An Introduction. Technical Report MS-CIS-04-21, Department of Computer and Information Science, University of Pennsylvania.
  • Wang, L., W. P. Sousa, and P. Gong. 2004. “Integration of Object-Based and Pixel-Based Classification for Mapping Mangroves with IKONOS Imagery.” International Journal of Remote Sensing 25 (24): 5655–5668. doi:10.1080/014311602331291215.
  • Wardlow, B. D., and K. Callahan. 2014. “A Multi-Scale Accuracy Assessment of the MODIS Irrigated Agriculture Data-Set (Mirad) for the State of Nebraska, USA.” GIScience & Remote Sensing 51 (5): 575–592. doi:10.1080/15481603.2014.952546.
  • Wawrzaszek, A., M. Krupinski, and S. Aleksandrowicz. 2013. “Fractal and Multifractal Characteristics of Very High Resolution Satellite Images.” Paper presented at the 2013 IEEE International Geoscience and Remote Sensing Symposium (IGARSS), July 21–26, Melbourne.
  • Whiteside, T. G., S. W. Maier, and G. S. Boggs. 2014. “Area-Based and Location-Based Validation of Classified Image Objects.” International Journal of Applied Earth Observation and Geoinformation 28: 117–130. doi:10.1016/j.jag.2013.11.009.
  • Woodcock, C. E., and A. H.Strahler. 1987. “The Factor of Scale in Remote Sensing.” Remote Sensing of Environment 21 (3): 311–332. doi:10.1016/0034-4257(87)90015-0.
  • Wu, C. S., C. B. Deng, and X. P. Jia. 2014. “Spatially Constrained Multiple Endmember Spectral Mixture Analysis for Quantifying Subpixel Urban Impervious Surfaces.” IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing 7 (6): 1976–1984. doi:10.1109/jstars.2014.2318018.
  • Xiao, J. F., and A. Moody. 2005. “A Comparison of Methods for Estimating Fractional Green Vegetation Cover within A Desert-To-Upland Transition Zone in Central New Mexico, USA.” Remote Sensing of Environment 98 (2–3): 237–250. doi:10.1016/j.rse.2005.07.011.
  • Xu, M. M., B. Du, and L. P. Zhang. 2014. “Spatial-Spectral Information Based Abundance-Constrained Endmember Extraction Methods.” IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing 7 (6): 2004–2015. doi:10.1109/jstars.2013.2268661.
  • Xu, M. M., L. P. Zhang, and B. Du. 2015. “An Image-Based Endmember Bundle Extraction Algorithm Using Both Spatial and Spectral Information.” IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing 8 (6): 2607–2617. doi:10.1109/jstars.2014.2373491.
  • Yedidia, J. S., W. T. Freeman, and Y. Weiss. 2003. “Understanding Belief Propagation and Its Generalizations.” Exploring Artificial Intelligence in the New Millennium 8: 236–239.
  • Yue, P., L. Di, Y. Wei, and W. Han. 2013. “Intelligent Services for Discovery of Complex Geospatial Features from Remote Sensing Imagery.” ISPRS Journal of Photogrammetry and Remote Sensing 83: 151–164. doi:10.1016/j.isprsjprs.2013.02.015.
  • Zare, A. 2011. “Spatial-Spectral Unmixing Using Fuzzy Local Information.” Paper presented at the 2011 IEEE International on Geoscience and Remote Sensing Symposium (IGARSS), July 24–29, Vancouver, BC.
  • Zare, A., O. Bchir, H. Frigui, and P. Gader. 2010. “Spatially-Smooth Piece-Wise Convex Endmember Detection.” Paper Presented at the 2010 2nd Workshop on Hyperspectral Image and Signal Processing: Evolution in Remote Sensing (WHISPERS), June 14–16, Reykjavik.
  • Zare, A., P. Gader, O. Bchir, and H. Frigui. 2013. “Piecewise Convex Multiple-Model Endmember Detection and Spectral Unmixing.” IEEE Transactions on Geoscience and Remote Sensing 51 (5): 2853–2862. doi:10.1109/tgrs.2012.2219058.
  • Zare, A., and K. C. Ho. 2014. “Endmember Variability in Hyperspectral Analysis: Addressing Spectral Variability During Spectral Unmixing.” IEEE Signal Processing Magazine 31 (1): 95–104. doi:10.1109/msp.2013.2279177.
  • Zhan, Q., M. Molenaar, K. Tempfli, and W. Shi. 2005. “Quality Assessment for Geo‐Spatial Objects Derived from Remotely Sensed Data.” International Journal of Remote Sensing 26 (14): 2953–2974. doi:10.1080/01431160500057764.
  • Zhang, B., S. Li, X.Jia, L. Gao, and M. Peng. 2011. “Adaptive Markov Random Field Approach for Classification of Hyperspectral Imagery.” IEEE Geoscience and Remote Sensing Letters8 (5): 973–977. doi:10.1109/LGRS.2011.2145353.
  • Zhang, C. Y., H. Cooper, D. Selch, X. L. Meng, F. Qiu, S. W. Myint, C. Roberts, and Z. X. Xie. 2014a. “Mapping Urban Land Cover Types Using Object-Based Multiple Endmember Spectral Mixture Analysis.” Remote Sensing Letters 5 (6): 521–529. doi:10.1080/2150704x.2014.930197.
  • Zhang, G., and X. Jia. 2012. “Simplified Conditional Random Fields with Class Boundary Constraint for Spectral-Spatial Based Remote Sensing Image Classification.” IEEE Geoscience and Remote Sensing Letters 9 (5): 856–860. doi:10.1109/LGRS.2012.2186279.
  • Zhang, H., J. Li, Y. Huang, and L. Zhang. 2014b. “A Nonlocal Weighted Joint Sparse Representation Classification Method for Hyperspectral Imagery.” IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing 7 (6): 2056–2065. doi:10.1109/JSTARS.2013.2264720.
  • Zhang, J., B. Rivard, and D. M. Rogge. 2008. “The Successive Projection Algorithm (SPA), an Algorithm with a Spatial Constraint for the Automatic Search of Endmembers in Hyperspectral Data.” Sensors 8 (2): 1321–1342. doi:10.3390/s8021321.
  • Zhang, J. S., C. Y. He, Y. Y. Zhou, S. Zhu, and G. Y. Shuai. 2014c. “Prior-Knowledge-Based Spectral Mixture Analysis for Impervious Surface Mapping.” International Journal of Applied Earth Observation and Geoinformation 28: 201–210. doi:10.1016/j.jag.2013.12.001.
  • Zhang, L., X. Huang, B. Huang, and P. Li. 2006. “A Pixel Shape Index Coupled with Spectral Information for Classification of High Spatial Resolution Remotely Sensed Imagery.” IEEE Transactions on Geoscience and Remote Sensing 44 (10): 2950–2961. doi:10.1109/TGRS.2006.876704.
  • Zhang, X. L., X. Z. Feng, P. F. Xiao, G. J. He, and L. J. Zhu. 2015a. “Segmentation Quality Evaluation Using Region-Based Precision and Recall Measures for Remote Sensing Images.” ISPRS Journal of Photogrammetry and Remote Sensing 102: 73–84. doi:10.1016/j.isprsjprs.2015.01.009.
  • Zhang, Z., C. Liu, J. Luo, Z. Shen, and Z. Shao. 2015b. “Applying Spectral Mixture Analysis for Large-Scale Sub-Pixel Impervious Cover Estimation Based on Neighbourhood-Specific Endmember Signature Generation.” Remote Sensing Letters 6 (1): 1–10. doi:10.1080/2150704X.2014.996677.
  • Zhao, J., Y. Zhong, and L. Zhang. 2015. “Detail-Preserving Smoothing Classifier Based on Conditional Random Fields for High Spatial Resolution Remote Sensing Imagery.” IEEE Transactions on Geoscience and Remote Sensing 53 (5): 2440–2452. doi:10.1109/TGRS.2014.2360100.
  • Zhao, Y. Y., P. Gong, L. Yu, L. Y. Hu, X. Y. Li, C. C. Li, H. Y. Zhang, et al.2014. “Towards a Common Validation Sample Set for Global Land-Cover Mapping.” International Journal of Remote Sensing 35 (13): 4795–4814. doi:10.1080/01431161.2014.930202.
  • Zhen, Z., L. J. Quackenbush, S. V. Stehman, and L. J. Zhang. 2013. “Impact of Training and Validation Sample Selection on Classification Accuracy and Accuracy Assessment When Using Reference Polygons in Object-Based Classification.” International Journal of Remote Sensing 34 (19): 6914–6930. doi:10.1080/01431161.2013.810822.
  • Zhong, C., C. Z. Wang, and C. S. Wu. 2015. “MODIS-Based Fractional Crop Mapping in the US Midwest with Spatially Constrained Phenological Mixture Analysis.” Remote Sensing 7 (1): 512–529. doi:10.3390/rs70100512.
  • Zhong, P., and R. Wang. 2010. “Learning Conditional Random Fields for Classification of Hyperspectral Images.” IEEE Transactions on Image Processing 19 (7): 1890–1907. doi:10.1109/TIP.2010.2045034.
  • Zhong, P., and R. Wang. 2011. “Modeling and Classifying Hyperspectral Imagery by Crfs with Sparse Higher Order Potentials.” IEEE Transactions on Geoscience and Remote Sensing 49 (2): 688–705. doi:10.1109/TGRS.2010.2059706.
  • Zhong, Y., X. Lin, and L. Zhang. 2014. “A Support Vector Conditional Random Fields Classifier with A Mahalanobis Distance Boundary Constraint for High Spatial Resolution Remote Sensing Imagery.” IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing 7 (4): 1314–1330. doi:10.1109/JSTARS.4609443.
  • Zhong, Y., J. Zhao, and L. Zhang. 2014. “A Hybrid Object-Oriented Conditional Random Field Classification Framework for High Spatial Resolution Remote Sensing Imagery.” IEEE Transactions on Geoscience and Remote Sensing 52 (11): 7023–7037. doi:10.1109/TGRS.2014.2306692.
  • Zhu, C., and X. Yang. 1998. “Study of Remote Sensing Image Texture Analysis and Classification Using Wavelet.” International Journal of Remote Sensing 19 (16): 3197–3203. doi:10.1080/014311698214262.
  • Zhu, F. Y., Y. Wang, S. M. Xiang, B. Fan, and C. H. Pan. 2014. “Structured Sparse Method for Hyperspectral Unmixing.” ISPRS Journal of Photogrammetry and Remote Sensing 88: 101–118. doi:10.1016/j.isprsjprs.2013.11.014.
  • Zortea, M., and A. Plaza. 2009. “Spatial Preprocessing for Endmember Extraction.” IEEE Transactions on Geoscience and Remote Sensing 47 (8): 2679–2693. doi:10.1109/tgrs.2009.2014945.