15,937
Views
25
CrossRef citations to date
0
Altmetric
Research Article

Deep learning for change detection in remote sensing: a review

ORCID Icon, , ORCID Icon, ORCID Icon, ORCID Icon, ORCID Icon & show all
Pages 262-288 | Received 19 Aug 2021, Accepted 30 May 2022, Published online: 01 Jul 2022

ABSTRACT

A large number of publications have incorporated deep learning in the process of remote sensing change detection. In these Deep Learning Change Detection (DLCD) publications, deep learning methods have demonstrated their superiority over conventional change detection methods. However, the theoretical underpinnings of why deep learning improves the performance of change detection remain unresolved. As of today, few in-depth reviews have investigated the mechanisms of DLCD. Without such a review, five critical questions remain unclear. Does DLCD provide improved information representation for change detection? If so, how? How to select an appropriate DLCD method and why? How much does each type of change benefits from DLCD in terms of its performance? What are the major limitations of existing DLCD methods and what are the prospects for DLCD? To address these five questions, we reviewed according to the following strategies. We grouped the DLCD information assemblages into the four unique dimensions of remote sensing: spectral, spatial, temporal, and multi-sensor. For the extraction of information in each dimension, the difference between DLCD and conventional change detection methods was compared. We proposed a taxonomy of existing DLCD methods by dividing them into two distinctive pools: separate and coupled models. Their advantages, limitations, applicability, and performance were thoroughly investigated and explicitly presented. We examined the variations in performance between DLCD and conventional change detection. We depicted two limitations of DLCD, i.e. training sample and hardware and software dilemmas. Based on these analyses, we identified directions for future developments. As a result of our review, we found that DLCD’s advantages over conventional change detection can be attributed to three factors: improved information representation; improved change detection methods; and performance enhancements. DLCD has to surpass the limitations with regard to training samples and computing infrastructure. We envision this review can boost developments of deep learning in change detection applications.

1. Introduction

Deep learning is often treated as a black box; thus, the theoretical underpinnings of deep learning mechanisms that improve change detection in remote sensing are little understood. In this paper, we argue that change detection performance improvements stemming from deep learning result from improved information representation and more effective models than those found in statistical, and machine learning methods, more commonly deployed. A comparative review of the related literature shows that Deep Learning Change Detection (DLCD) provides semantic information neglected by other methods, and flexible network configurations that can improve change detection methods and performance. Semantic information, improved change detection methods, and performance enhancements facilitate improved change detection outcomes in remote sensing applications.

1.1. Background of change detection

Change detection aims to discern both subtle and abrupt alternation in a given area as manifested through remote sensing images acquired at different times (Singh Citation1989). It provides a means to study land use/cover change (Weng Citation2002), biodiversity (Newbold et al. Citation2015), the urbanization process (Han et al. Citation2017a), disaster detection (Saito et al. Citation2004), and other environmental changes. Conventionally, two types of changes are often examined (Lu et al. Citation2004): 1) binary changes, and 2) from-to changes. The binary changes only focus on where change has been incurred or not. From-to change detection not only detects variation over time but also the specific types of change e.g. from buildings to vegetation. Lately, a new type of change has been studied: multi-class changes, i.e. clustering the detected changes into different groups regardless of the land cover types (Saha, Bovolo, and Bruzzone Citation2019).

Substantial research has been carried out to recognize these three types of change. The definition of change detection was first introduced by Singh (Citation1989) but more recently, Tewkesbury et al. (Citation2015) summarized all the change detection methods in six categories: 1) Layer arithmetic methods (Howarth, and Wickware Citation1981), 2) Post-Classification Change Methods (PCCMs) (Silván-Cárdenas, and Wang Citation2014; Yuan et al. Citation2005), 3) Direct Classification Methods (DCMs) (Bovolo, Bruzzone, and Marconcini Citation2008), 4) Transformation methods (Gong Citation1993), 5) Change Vector Analysis (CVA) methods (Chen et al. Citation2003; Johnson, and Kasischke Citation1998), and 6) Hybrid change detection methods (Healey et al. Citation2018; McDermid et al. Citation2008). Among these six categories, the layer arithmetic and transformation methods first obtain a Difference Image (DI) and then select a threshold to discriminate the DI to obtain binary changes. These methods are easy to implement but cannot detect from-to changes because of the information loss in the process of obtaining the DI.

Unlike these two categories, PCCMs and DCMs are both to classify multi-date images to detect changes. The difference between these two categories is that PCCMs conduct change analysis after obtaining the independent classification map of each image, whereas DCMs directly classify stacked features obtained from multi-date images to detect changes. As a result, PCCMs can be used for multi-sensor images, and detect both binary changes and from-to changes. However, errors stemming from the classification maps are compounded in the final change map, thus reducing the accuracy of the final change detection result (Chan, Chan, and Yeh Citation2001; Dai, and Khorram Citation1999; Lillesand, Kiefer, and Chipman Citation2015). On the other hand, DCMs can overcome the error propagation problem, but requires training samples of all change types. To simultaneously mitigate the error propagation problem and the sampling difficulty, the fifth category, CVA, was developed. CVA constructs change vectors, and then utilizes their magnitudes and directions to detect changes. This method can detect binary changes, but its ability to detect from-to changes is limited (Tewkesbury et al. Citation2015). To combine the merits of these five categories, the last category, hybrid change detection methods, emerged. However, this category simultaneously inherits the weaknesses of the change detection methods used.

The big data era has further complicated the problems associated with conventional change detection methods (Reichstein et al. Citation2019). This is manifested in the following three ways: 1) Data volume of change detection is unprecedentedly large and quickly expanding, compounding the weaknesses. For example, an error will be further propagated when more data layers are engaged; the required number of training samples is escalating; from-to changes cannot be easily uncovered from the time series data. 2) The data sources of change detection are more diverse than ever. Contemporary data are comprised of multiple spectrums of platforms (satellite and aerial), sensors (passive and active), spatial resolution (250 m to sub-meter), and spectral resolution (multispectral and hyperspectral). This diverse source makes information expression not consistent. What’s more, there unavoidably exists a problem of spectral variability in more hyperspectral imagery (Hong et al. Citation2018), which leads to inaccurate change detection results. 3) The data tend to suffer from various degradation, noise effects, or variabilities in the process of imaging due to atmospheric conditions, illumination, viewing angles, soil moisture, etc. These factors are random and difficult to be considered precisely in traditional change detection methods. Therefore, how address these three issues have become a pressing challenge before change detection can be working for the big remote sensing data.

1.2. Deep learning methods used in change detection

The evolution of deep learning has demonstrated immense potential for addressing various change detection challenges in remote sensing. Deep learning is a particular kind of machine learning method based on artificial neural networks with representation learning (LeCun, Bengio, and Hinton Citation2015). Compared with other machine learning methods, the deep learning model achieves great power and flexibility by representing the world as a nested hierarchy of concepts, with each concept defined about simpler concepts, and more abstract representations computed in terms of less abstract ones (LeCun, Bengio, and Hinton Citation2015).

As of today, deep learning has been intensively used in the remote sensing field (Ma et al. Citation2019a), such as image fusion (Shao, and Cai Citation2018), image registration (Wang et al. Citation2018b), image matching (Chen, Rottensteiner, and Heipke Citation2021; Zhang et al. Citation2020; Cheng et al. Citation2020), change detection (Zhu et al. Citation2018), land use and cover classification (Hong et al. Citation2021, Citation2020b, Citation2020a; Li et al. Citation2022; Zhou et al. Citation2022), semantic segmentation (Kemker, Salvaggio, and Kanan Citation2018), and object-based image analysis (Liu, Yang, and Lunga Citation2021). Specifically, a new sub-field, known as DLCD, has emerged when deep learning methods are employed to detect changes in multi-date remote sensing images.

Compared to other topics in remote sensing, DLCD is still in its infancy. We carried out a search of all the published papers that involves DLCD using the search query “deep learning AND change detection” in the Scopus database, similar to Ma et al. (Citation2019a). Altogether 80 publications () have witnessed the early-stage developments of DLCD-relevant applications between 2013 and 2019, including peer-reviewed articles and conference papers.

Figure 1. Publication numbers per year of DLCD.

Figure 1. Publication numbers per year of DLCD.

There is not a lot of DLCD literature as compared with the scene classification and object detection, land use and cover classification, and semantic segmentation tasks because DLCD faces additional challenges. While segmentation and classification tasks work on images at a single time point, change detection works simultaneously on images at multiple time points. Therefore, the problems and limitations (e.g. noise) associated with single-date image information extraction are multiplied in DLCD. Furthermore, the multi-date images may contain different information expressions (e.g. inconsistent radiometric and geometric information, different spatial resolution, and different sensor), which makes DLCD even more difficult.

Another particular difficulty in DLCD is that change is often minimal in the images. While segmentation and classification tasks extract the bulk of the relevant information, change detection aims to find the minority pixels that have experienced changes. Because changes only occur in limited amounts (<50%) in general, the change information is easily confused with noise. The quantity of the related DLCD publications, however, is increasing rapidly () due to the need of change detection applications.

Among these studies, the most popular Deep Learning Neural Networks (DLNNs) (i.e. deep learning models) used in change detection are Convolutional Neural Networks (CNNs), Auto-Encoders (AEs), Deep Belief Networks (DBNs), Recurrent Neural Networks (RNNs), and Generative Adversarial Networks (GANs). DLNNs may each contain multiple specific models. For example, CNNs can be implemented using ResNet, AlexNet, DenseNet, etc. To discuss the structures of these specific models are beyond the scope of this review; thus, these specific models are not included. For each type of DLNNs, lists the first work that proposed the DLNN (deep learning reference) and the first work that used the DLNN in change detection (DLCD reference). To further understand the structure of specific DLNNs or how they are used in change detection, readers are referred to these publications.

Table 1. The first proposed reference and DLCD reference for each DLNN.

1.3. Unsolved problems in DLCD reviews

Until now, most existing studies reviewing deep learning are general reviews concerning the algorithm development of deep learning and remote sensing specific applications (Zhang, Zhang, and Du Citation2016; LeCun, Bengio, and Hinton Citation2015; Ball, Anderson, and Chan Citation2018; Ma et al. Citation2019a; Yuan et al. Citation2020; Zhu et al. Citation2017). Zhu et al. (Citation2017) and Ma et al. (Citation2019a) have reviewed applications and technology of deep learning remote sensing from preprocessing and mapping. Yuan et al. (Citation2020) have compared the use of the traditional neural network and deep learning methods to advance the environmental remote sensing process. There are two review papers for DLCD, which provided a technical review of the advance of deep learning for change detection (Shi et al. Citation2020; Khelifi, and Mignotte Citation2020). These two reviews are from the perspectives of the implementation process, technical methods, and data sources, providing a good starting point for beginners to understand DLCD. However, the theoretical underpinning of why deep learning improves the performance of change detection as compared with conventional change detection remains unresolved. Such knowledge could enable refinements in the existing deep learning methods and thus supersede current limitations such as data pollution and opaque network configurations. Specifically, to unveil the black box of DLCD, the following questions have to be investigated.

1) Does DLCD provide improved information representation for change detection? If so, how?

The first step in DLCD begins with data input. In general, four different types of inputs are engaged in a typical DLCD study, i.e. spectral, spatial, temporal, and multi-source information. It is critical to understand whether DLCD has a more powerful capability to represent the four types of information compared to the conventional change detection methods. If so, how such information is represented. Without such understanding, it is a challenge to further refine DLCD methods to achieve the best tradeoff that can maximize the authentic information carried over in the input and suppressed the noise at the same time.

2) How to select an appropriate DLCD method and why?

Methodology plays an important role in change detection. The introduction of deep learning network architectures opens up a new avenue for change detection. Since 2013, a variety of DLCD methods has emerged. Each DLCD method has its respective combination of network layers, training samples’ requirements, and applicability. These characteristics result in different performances. A systematical comparison of different advantages, disadvantages, and performance of DLCD methods would give methodological guidance in future change detection applications. However, such a comparison is not available. Without such knowledge, it is a challenge to select DLCD methods suitable for actual applications.

3) How much does each type of change benefits from DLCD in terms of their performance?

There are three types of change detection results including binary changes, multiclass changes, and from-to changes. The complexity of information processing increases from binary change detection to from-to change detection. As a result, the performance improvement from conventional change detection to DLCD is different regarding these three types of changes. Such difference is even more varied depending on the application. However, this difference has not been reviewed systematically. Given the fact that deep learning has a larger computation burden, a better understanding of the difference can help to measure the accuracy vs. computation burden tradeoff when selecting deep learning or conventional methods in different change detection applications.

4) What are the major limitations of existing DLCD methods?

The incorporation of deep learning introduces two major limitations including training sample and hardware and software dilemmas except greatly benefiting change detection. The excellent performance of typical deep learning models relies on extremely large labeled training sample sets (Gong et al. Citation2019b; Wang et al. Citation2018a) but geospatial systems usually have particularly limited labeled training samples available (Ball, Anderson, and Chan Citation2018; De et al. Citation2017; Reichstein et al. Citation2019), especially change detection. Deep learning places high demands on computational power in hardware and software. These pose two big challenges for DLCD, as no research is available. The lack of such research may become a major hurdle for further applying deep learning in change detection.

5) What are the prospects for DLCD?

Based on the analysis of the benefits and limitations of DLCD, we can outline prospects. Although there are some discussions of future directions in the DLCD literature, these articles have tended to focus on the disadvantages of methods themselves, or implementation processes, and so there is no overview of directions for future research based on the theoretical underpinnings. This paper addresses this gap.

1.4. Structure of this review

To answer the five questions, our review made five major contributions, as follows:

  1. We compare the difference in how spectral, spatial, temporal, and multi-sensor information are represented in DLCD and conventional change detection methods.

  2. We propose a taxonomy for DLCD methods by dividing them into two distinctive pools: separate and coupled DLNNs, based on which, a thorough analysis of their advantages, limitations, applicability, and performance is investigated.

  3. We examine the holistic accuracy between DLCD and conventional change detection methods by adopting a box plot to analyze four major land-use types: Urban, Water, Vegetation, and Hazards for the three primary change types: i.e. binary, multiclass, and from-to.

  4. We discuss two limitations of DLCD including training sample and hardware and software dilemmas.

  5. We identify four directions for future development.

2. Improved information representation

The success of change detection depends on maximizing real change. Although both conventional change detection and DLCD aims to achieve this goal, the key difference between the two groups of methods is in the execution. DLCD is optimized for spectral, spatial, temporal, and multi-sensor information representation. In this section, we will compare how these kinds of information are extracted by DLCD as opposed to the conventional change detection methods.

2.1. Spectral information

DLCD is optimized for abstract spectral information representation (such as spectral curve) compared with the conventional change detection methods. Traditional thresholding-based, transformation, spectral analysis, and classification-based methods (Liu et al. Citation2014) can extract low-dimension spectral information (such as the mean value(R, G, B), the standard variance(R, G, B), the brightness, and the maximum difference). However, they neglect abstract spectral information, resulting in inaccurate recognition of complicated land use and land cover changes.

DLCD adopts a different strategy to discern real changes based on multi-scale spectral information differences, as shown in . outlines a basic scenario of change detection based on spectral information. T1 and T2 represent the reflectance at two different times (i.e. before the change and after change). Two rectangular boxes represent the multi-scale spectral information difference between T1 and T2. Typical change detection detects changes based on the spectral information difference presented between T1 and T2 reflectance curves.

Figure 2. The usage of spectral information in DLCD.

Figure 2. The usage of spectral information in DLCD.

From , one example of such multi-scale spectral information difference can be understood as conducting an average of the original spectral information curve (e.g. T1 and T2) with a different number of spectral bands. Once an average curve is derived, the difference between T1 and T2 is calculated. Then all the changes at disparate scales will be accumulated before an overall conclusion of change can be derived. Technically, such multi-scale integration is implemented using a large number of neurons and a complex network structure for deep learning. The advantages of deep learning become more prominent as the spectral band information increases. DLCD methods can directly handle the high-dimensional spectral information and more effectively learn available multi-scale spectral information. This information includes detailed spectral information at a low-level scale and semantic information at a high-level scale (Wang et al. Citation2019; Li, Yuan, and Wang Citation2019).

2.2. Spatial information

DLCD is optimized for multi-scale spatial information extraction as compared with the conventional change detection methods. This spatial information includes texture, location information at low scale, and semantic information at a high scale, and the most commonly used conventional change detection methods for extracting spatial information are object-based (Zhou, Yu, and Qin Citation2014; Chehata et al. Citation2014; Tang, Huang, and Zhang Citation2013), which can extract multi-scale and hierarchical spatial information features. However, the extraction accuracy is heavily dependent on the accuracy of the initial segmentation (Cao et al. Citation2014; Tewkesbury et al. Citation2015; Hussain et al. Citation2013). Another common method is to extract spatial features based on the fixed and limited window such as the gray level co-occurrence matrix (Murray, Lucieer, and Williams Citation2010). Thus, these extracted spatial features are limited-scale features, are not complete and systematic, and thus are not robust enough to detect complicated urban land use and land cover changes.

In contrast to the conventional change detection methods, DLCD utilizes multi-scale spatial information differences to discern real changes, presented in . outlines an example of change detection based on spatial information. A typical change detection approach detects changes based on the spatial information difference presented between T1 and T2.

Figure 3. An example of the usage of spatial information in DLCD.

Figure 3. An example of the usage of spatial information in DLCD.

From , one example of such multi-scale spatial information difference can be understood as extracting the spatial features (e.g. T1 and T2) utilizing CNN. CNN utilizes the convolution kernels of different sizes (e.g. 5 × 5) to extract spatial features from a 32 × 32 patch in T1 and T2. Once the spatial features are derived, the spatial information difference between T1 and T2 can be calculated. All the changes at disparate scales will be accumulated before an overall change determination can be derived. Technically, such multi-scale integration is implemented using multi-layer convolutions. Unlike conventional change detection methods, DLCD can automatically extract multi-scale spatial information (Amit, and Aoki Citation2017; El Amin, Liu, and Wang Citation2016; Khan et al. Citation2017).

2.3. Temporal information

DLCD is optimized for extracting nonlinear temporal information compared with the conventional change detection methods. Conventionally, time-series change detection utilizes statistical metrics such as individual ranks, means, and regression slopes for spectral bands and vegetation indices as the input (Shao, and Liu Citation2014). These are generally calculated either based on time-sequential reflectance or reflectance rank (Zhu Citation2017; Shao et al. Citation2019a); however, time-series change detection uses limited and linear statistical features to depict continuous temporal information, neglecting other types of temporal information.

DLCD adopts a recurrent method (Section 3.2.3) to discern real changes based on the differences in the time-series curves, as shown in . outlines a basic scenario of change detection based on temporal information. The blue curve represents Pixel 1 (Change) and the yellow curve represents Pixel 2 (No change). A typical time-series change detection detects changes based on the temporal feature differences from the time-series curves between T1, T2, … , and TN for different pixels.

Figure 4. The usage of temporal information in DLCD.

Figure 4. The usage of temporal information in DLCD.

From , for Pixel 1 and Pixel 2, the reflectance {X1, X2, … , and XN} in T1, T2, … , and TN is the input of the recurrent method. The recurrent method learns temporal information by building recurrent connections between T1, T2, … , and TN. Thus, the output feature information of TN depends on input XN in TN and feature information in TN-1. Once the temporal feature in TN for Pixels 1 and 2 is derived, the feature difference between Pixel 1 and Pixel 2 is calculated. An overall conclusion of change can be derived. Technically, such a recurrent method is implemented using the recurrent connections between the neural activations of RNN at consecutive time steps. In contrast to traditional change detection methods, the DLCD can learn detailed temporal information as well as nonlinear temporal information automatically, supporting the detection of complicated urban land-use change (Lyu, Lu, and Mou Citation2016; Mou, Bruzzone, and Zhu Citation2018) and vegetation phenological patterns (Song et al. Citation2018). The advantages of DLCD become more pronounced as the time series expands.

2.4. Multi-sensor information (multi-source/multi-spatial)

DLCD is optimized for extracting multi-scale features compared with the conventional change detection methods. Change detection methods for multi-sensor images can be divided into two categories: 1) comparative analysis of independently produced classifications for different sensors and 2) simultaneous analysis of multi-temporal data (Singh Citation1989; Zhang et al. Citation2019). The former conducts multi-sensor images classification separately for change detection analysis. DLCD also has this type of method, i.e. the post-classification change method (Section 3.1.1.1), which extracts high-dimensional feature representations to replace limited and low-level feature extractions (Iino et al. Citation2018; Nemoto et al. Citation2017). This method obtains higher accuracy than the traditional post-classification methods (Iino et al. Citation2018; Nemoto et al. Citation2017).

The simultaneous analysis of multi-temporal data methods includes the copula theory (Mercier, Moser, and Serpico Citation2008), manifold learning strategy (Prendes et al. Citation2015b), kernel canonical correlation analysis (Volpi, Camps-Valls, and Tuia Citation2015), and Bayesian nonparametric model associated with a Markov random field (Prendes et al. Citation2015a). However, these methods utilize the limited features to find the relationship between the unchanged areas in multi-sensor remote sensing images and they do not consider the influence of changed areas when identifying this relationship.

DLCD adopts a mapping transformation method (Section 3.1.5.2.2) to discern authentic changes based on different sensor information differences between T1 and T2, as shown in . outlines an example of the usage of multi-sensor (multi-source and multi-spatial) information in DLCD. A typical multi-sensor change detection is how to detect changes from T1 and T2 for different sensors.

Figure 5. An example of the usage of multi-sensor information in DLCD.

Figure 5. An example of the usage of multi-sensor information in DLCD.

From , one example of such multi-sensor information difference can be understood as detecting changes between Landsat (T1) and sentinel 1 (T2). The mapping transformation method transforms T1 and T2 image information into a multi-scale feature space to obtain features similar to those found in the other image or finds the relationship between two image features. Once the similar features or the relationships between features in two images are derived, the difference between T1 and T2 can be calculated, and an overall determination of change can be derived. Technically, the mapping transformation method is implemented by building transformation neural networks. Unlike traditional change detection methods, the DLCD method learns multi-scale features and transforms inconsistent information into consistent information. In addition, it can enlarge the distance between changed pixels as the constraint rule to find more accurate relationships between information transformations.

3. Improved change detection methods

DLCD improves change detection methods as well as information representation for change detection. To provide methodological guidance for future applications, we present a taxonomy of these methods and provide a systematic comparison of their advantages, limitations, applicability, and performance. DLCD methods are separated into two categories at the first level as either separate or coupled DLNNs. At the second level, seven sub-categories are included. The separate DLNNs contain three sub-categories: PCCM, Differencing Methods (DM), and DCM. The coupled DLNNs contain four sub-categories: Differencing Neural Network Method (DNNM), Mapping Transformation Method (MTM), Recurrent Method (RM), and Adversarial Method (AM). A diagram of separate DLNNs and coupled DLNNs is shown in . The frameworks are provided for each sub-category.

Figure 6. The structure of DLCD methods. Yellow boxes represent DLNNs while blue boxes denote the interaction of multi-date information. (a) Post-classification change method (PCCM). (b) Differencing method (DM). (c) Direct classification method (DCM). (d) Differencing neural network method (DNNM). (e) Mapping transformation method (MTM). (f) Recurrent method (RM). (g) Adversarial method (AM).

Figure 6. The structure of DLCD methods. Yellow boxes represent DLNNs while blue boxes denote the interaction of multi-date information. (a) Post-classification change method (PCCM). (b) Differencing method (DM). (c) Direct classification method (DCM). (d) Differencing neural network method (DNNM). (e) Mapping transformation method (MTM). (f) Recurrent method (RM). (g) Adversarial method (AM).

From , change detection requires the interaction between multi-date information. The main difference between separate DLNNs and coupled DLNNs is clearly shown as the relationship between DLNNs (yellow boxes) and the interaction (blue boxes). In separate DLNNs, DLNNs are used to generate deep features before interaction or after interaction between multi-date image information, but DLNNs do not directly interact. Coupled DLNNs, in contrast, multi-date interaction occurs between DLNNs. The taxonomy of DLCD methods is summarized in . The difference, definition, advantages, limitations, and application examples are provided for each type of DLCD.

Table 2. The taxonomy, difference, definition, advantages, limitations, and applications of DLCD methods.

From , separate DLNNs are adapted from conventional change detection methods by refining image features with DLNNs. They share the same principles as the corresponding conventional change detection methods. The coupled DLNNs, on the other hand, are novel strategies where multiple DLNNs interact to maximize the change information. Unlike separate DLNNs, coupled DLNNs require more knowledge about deep learning and cannot be directly compared to conventional change detection methods. A more detailed discussion of each sub-category is provided in the following sections.

3.1. Separate DLNNs

In this sub-section, we introduce three types of methods that do not allow DLNNs to interact but only use DLNNs to extract deep features, i.e. separate DLNNs (). They are adapted from traditional PCM, layer arithmetic and transformation, and traditional DCM, respectively.

3.1.1. Post-classification change method

PCCM compares classification maps from different times to identify changes between them. PCCM includes three steps (). First, feature learning; two images are input into separate DLNNs to obtain deep feature representations of each image respectively. Secondly classification; the deep features are classified separately to obtain two classification maps. Third, is change analysis; the two classification maps are compared to obtain the final change map. The difference between this method and a traditional PCCM is that feature learning is through a deep learning model (Nemoto et al. Citation2017; Iino et al. Citation2018).

PCCM is easy to implement because the DLNN is only applied for feature learning but no effort is required for the configuration of the DLNN structure. Another advantage of PCCM is that it works on multi-sensor images and does not require radiometric normalization, because the changes are detected from two independent classification maps. In addition, PCCM can detect complete from-to changes. Nevertheless, similar to the traditional PCCM, the disadvantage is that the errors from two classification maps can accumulate in the final change detection map. To date, this method has only been used to detect changes in urban land use (Nemoto et al. Citation2017) ().

3.1.2. Differencing method

DM detects change by comparing the differences in image radiance (Li et al. Citation2017; Liu et al. Citation2016b; Xiao et al. Citation2018) or deep features (Gong et al. Citation2017b; Su et al. Citation2016; Xu et al. Citation2013), and then changes are detected with a supervised or unsupervised deep learning technique. If differences in image radiance are used in DM, we refer to it as image DM. Image DM has three steps ()-(1)). First, differencing; a DI is obtained. Second, is feature learning; DLNNs are built to extract the deep features of this DI. Third, classification; the deep features are input into a soft classifier to obtain the final change map. If differences in deep features are used in DM, we refer to it as feature DM. Feature DM also has three steps ()-(2)). First, is feature learning; two images are input into two DLNNs to obtain deep features. Second, differencing; different features are obtained. Third, classification; these features are input into a classifier to obtain the final change detection result.

The key step in DM is differencing, which aims to suppress unchanged information and highlight changed information. Although the techniques that have been used to obtain DI or difference features are limited, more are available. Similar to traditional layer arithmetic, operations such as subtracting (Arabi, Karoui, and Djerriri Citation2018), log rationing (Gong, Yang, and Zhang Citation2017), and log-mean rationing (Li et al. Citation2017) can be applied. Techniques such as Principal Component Analysis (PCA) can serve a similar purpose. For example, El Amin, Liu, and Wang (Citation2017) applied PCA to stacked high resolution images in order to identify changed features.

In summary, like the PCCM, DM is also easy to implement and the subsequent change detection result is easy to interpret. In this method, only one classification stage is required and identified changes are thematically labeled. Nevertheless, DM does not provide complete from-to changes. To date, this method has been widely used for urban land use (Xu et al. Citation2013), water and hazard (Zhao et al. Citation2014), and vegetation change detections (Li et al. Citation2017) ().

3.1.3. Direct classification method

DCM directly classifies stacked multi-date images or deep features for change detection. If images are stacked in DCM, we refer to it as image stacking DCM. Image stacking DCM includes three steps ()-(1)). Multi-date images are stacked and the stacked images are input into DLNN to learn the deep features. The resulting deep features are input into a soft classifier for change detection.

If deep features are stacked in DCM, we refer to it as feature stacking DCM. Feature stacking DCM also includes three steps ()-(2)). Two images are input into DLNNs to obtain deep feature representations of each image respectively. The two sets of deep features are stacked. The stacked features are input into a soft classifier to obtain the final change map.

In summary, like the other separate DLNNs (PCCM and DM), DCM is also easy to implement. Like traditional DCM, this method requires only one classification stage and can identify changes thematically (Tewkesbury et al. Citation2015). However, the disadvantage is difficult to construct training samples for from-to-change detection. To date, this method has been widely used for urban land use (Zhang et al. Citation2016a), water (Gao et al. Citation2019a), and hazard and vegetation change detections (Gong et al. Citation2015) ().

3.2. Coupled DLNNs

In this sub-section, we introduce four types of methods that not only allow DLNNs to interact but also use DLNNs to extract deep features, i.e. coupled DLNNs (). Compared to the separate DLNNs where DLNNs are adapted from the traditional change detection methods, four types of methods in the coupled DLNNs category are novel implementation strategies for specific change detection application scenarios. DNNM can highlight different information, so it is resistant to noise. MTM can deal with multi-source images and multi-spatial-resolution images. RM can be used in change detection applications that are time-sensitive (e.g. crop growth). AM can eliminate noise and generate high-quality information.

3.2.1. Differencing neural network method

DNNM builds a cost function between two DLNNs to highlight the difference in deep features for change detection. DNNM includes three steps (). Two images are input into DLNNs to pre-train two DLNN models. A cost function is used to adjust the parameters of the two DLNNs to generate deep features. With these deep features, the difference between changed pixels at T1 and T2 is enlarged. The two sets of deep features are input into a soft classifier to obtain the final change detection result.

The key step in DNNM is fine-tuning, i.e. the cost function, which aims to highlight the changed information and suppress the unchanged information. There are different methods to build a cost function. For example, Chen, Shi, and Gong (Citation2016) used the difference between bi-temporal deep features and an initial DI to design a cost function. Cao et al. (Citation2017) extended a backpropagation algorithm to build a cost function between two DBNs.

In summary, these methods enhance the difference in change areas and effectively suppress the noise as compared to traditional DI creation methods (Chu, Cao, and Hayat Citation2016). Compared to the methods in separate DLNNs, especially DM, the advantage of this method is that it uses an extra cost function to simultaneously adjust the parameters of two DLNNs, which makes the network parameters more accurate. However, the disadvantage is that because of this additional cost function, the structure of this method is complex. It is difficult to construct training sample sets for from-to change detection. To date, this method has been used for urban land use (Chu, Cao, and Hayat Citation2016), water and hazard (Chen, Shi, and Gong Citation2016), and vegetation change detections (Geng et al. Citation2017) ().

3.2.2. Mapping transformation method

MTM constructs a transformation function layer between inconsistent deep feature representations for change detection. For multi-spatial resolution or multi-source remote sensing images, the direct comparison between pixel-pair or feature-pair is meaningless (Zhan et al. Citation2018; Zhang et al. Citation2016b). Therefore, MTM was proposed to explore the inner relationships between multi-sensor data.

MTM includes four steps (). Two images are input into DLNNs to obtain deep feature representations of each image, respectively. A transformation function between the two sets of deep features is constructed. The similarity of the transformed features of the two images is calculated, and similarity features are clustered to obtain the final change map.

The key step in MTM is transformation. The transformation function can be constructed based on different principles. A transformation function can be built based on the principle that unchanged pixels at the same position in two input images have similar representations (Liu et al. Citation2016a; Zhang et al. Citation2016b). Conversely, the transformation function can be built by shrinking the difference between the paired features of unchanged positions while enlarging the difference between the paired features of changed positions (Zhan et al. Citation2018). The target of these transformation functions is to make incomparable information from different sensors comparable. For instance, Zhang et al. (Citation2016b) for the first time proposed a Stacked Denoised Auto-Encoder (SDAE) based MTM to detect changes between images of different resolutions. Liu et al. (Citation2018) established a novel MTM by using a convolutional coupling network to detect the change between optical and radar images. The success in these examples showed that this method is capable of detecting changes from multi-sensor images.

MTM is a powerful tool for transforming information from heterogeneous images into consistent information. However, like the DNNM, the disadvantage is that the additional transformation function makes its structure more complex. It is also difficult to construct training samples and provide complete from-to changes. To date, this method has been used for detecting changes in urban land use (Zhan et al. Citation2018), water and hazard (Zhang et al. Citation2016b), and vegetation applications (Su et al. Citation2017) ().

3.2.3. Recurrent method

RM uses recurrent connections between multi-date images to learn deep features and includes three steps (). First, is feature learning. DLNNs such as DBN, AE, and CNN are used to extract spectral or spatial features, while RNNs extracts temporal features. The temporal features are input into a soft classifier to obtain the final change detection result.

This method adds temporal features to traditional spectral or spatial feature methods. For instance, Lyu, Lu, and Mou (Citation2016) used a long short-term memory network to learn spectral and temporal features for change detection. Mou, Bruzzone, and Zhu (Citation2018) proposed a new RM by combining CNN and RNN to learn joint spectral-spatial-temporal features for change detection. These examples demonstrate that learning temporal features can be an effective way to detect change. In addition, it can learn the phenological characteristics of vegetation. For example, Song et al. (Citation2018) proposed an RM by using a 3D fully convolutional network and a convolutional long short-term memory network to learn phenological features for change detection. These examples show that extracting phenological features can effectively facilitate change detection.

As compared to other methods, RM with the help of RNN considers the temporal connections in multi-temporal change detection tasks. A disadvantage is that the combination of RNN and other DLNNs makes its structure complex. It is also difficult to construct training sample sets and provide complete from-to changes. To date, this method has been used for detecting changes in urban land use (Lyu, Lu, and Mou Citation2016) and vegetation applications (Song et al. Citation2018) ().

3.2.4. Adversarial method

AM plays generator neural networks and the discriminator neural networks against each other (i.e. GANs) to achieve change detection. AM includes three steps (). The generator is used to extract deep features for each image. The deep features are stacked (Gong et al. Citation2019b) or differenced (Gong et al. Citation2017a) to obtain a feature map. The discriminator is used to discriminate the feature map (e.g. generated DI) from the real data (e.g. real DI). A final feature map is obtained when the discriminator cannot distinguish the feature map from the real data. This classification method is used to classify a feature map to obtain a binary change detection map.

The generator is common DLNNs such as CNN (Gong et al. Citation2017a). Choosing which DLNN for the generator depends on the specific application (Radford, Metz, and Chintala Citation2015). Gong et al. (Citation2017a) used this AM based on GANs to generate a better DI for change detection, which has less noise than a real DI. Gong et al. (Citation2019b) used this method to generate training data and used these additional training data, label data, and unlabeled data to build a semi-supervised classifier for change detection. The success of these two examples shows that this method has a powerful ability for generating high-quality information.

Compared with other methods, the advantage of AM is that it can eliminate noise and generate high-quality information, which provides a new avenue for change detection tasks such as generating training samples and better change information. However, the disadvantage is the combination of generator and discriminator neural networks makes its structure complex. It is also difficult to provide complete from-to changes. To date, this method has been used for detecting changes in urban land use and water applications (Gong et al. Citation2017a) ().

3.3. Performance comparison

Performance comparisons of the overall accuracy of different DLCD methods, including DM, DCM, DNNM, MTM, RM, and AM were made using a boxplot, as shown in . The accuracy values for each DLCD method are from case studies in the respective DLCD references. Some of the referenced methods may use the same dataset. For example, the Ottawa dataset is used for evaluation studies of the difference (Li et al. Citation2017) and direct classification methods (Gao et al. Citation2017). We did not include PCCM in because the PCCM publications (Lyu, and Lu Citation2017; Cao, Dragićević, and Li Citation2019; Iino et al. Citation2018; Nemoto et al. Citation2017) did not report the overall change detection accuracy.

Figure 7. Distribution of overall accuracies for DLCD methods (differencing method (DM), direct classification method (DCM), differencing neural network method (DNNM), mapping transformation method (MTM), recurrent method (RM), and adversarial method (AM)).

Figure 7. Distribution of overall accuracies for DLCD methods (differencing method (DM), direct classification method (DCM), differencing neural network method (DNNM), mapping transformation method (MTM), recurrent method (RM), and adversarial method (AM)).

From , using the median of the accuracy values as an indicator of DLCD performance, DLCD methods can be ranked in descending order as DNNM, DCM, RM, DM, MTM, and AM. That DNNM ranks number one can be attributed to that DNNM is more resistant to noise than the other methods. AM can automatically generate training samples, so unsupervised and semi-supervised classifiers are often incorporated into this method. Thus, the median accuracy was slightly lower. In addition, the accuracy variability of the coupled DLNNs was much lower as compared with the separate DLNNs, as illustrated by the Inter-Quartile Range (IQR) within each DLCD method, i.e. the IQR using the coupled DLNNs is smaller than the IQR using separate DLNNs; this implies that the interaction between DLNNs can make change detection more robust.

4. Performance enhancements

DLCD optimized performance in three types of existing change detection (binary, multi-class, and from-to change detection), but how much each type of change detection in specific applications benefits from DLCD in terms of their performance is not yet clear. In the following, we review the difference in the overall improvement in accuracy between binary, multi-class, and from-to change detection for different applications. Due to space restrictions, not all potential applications and references are included. We explore the most popular applications including urban land use, water, hazard, and vegetation change detection.

4.1. Binary changes

The overall accuracy of binary change detection in urban land use, water, hazard, and vegetation applications using the conventional change detection and the DLCD methods is shown in . With regards to the four applications: urban land use, water, vegetation, and hazard, the overall accuracy values of conventional change detection and DLCD methods were extracted from case studies presented in representative references where both conventional change detection and DLCD methods are used.

Figure 8. Distribution of overall accuracies of binary changes for urban land use, water, hazard, and vegetation applications using the conventional change detection (CD) and DLCD methods.

Figure 8. Distribution of overall accuracies of binary changes for urban land use, water, hazard, and vegetation applications using the conventional change detection (CD) and DLCD methods.

From , the DLCD methods improved performance in comparison with conventional change detection methods in urban land use, water, hazard, and vegetation change detection applications. This can be evidenced by the fact that the overall median accuracy using the DLCD methods was higher than that using the conventional change detection methods for urban land use (92.55% to 96.07%), water (95.35% to 96.99%), hazard (96.91% to 98.09%), and vegetation (95.18% to 97.23%). The second conclusion is that the accuracy variability of the DLCD methods was lower than the conventional change detection methods. This can be evidenced by the respective IQR within each application, i.e. IQR using the DLCD methods is smaller than that using the conventional change detection methods. This is because DLCD yields a more precise feature representation for binary change detection.

In addition, the increase in the median overall accuracy for urban land use (increased by 3.52%) are more apparent than the increase in accuracy for water (increased by 1.64%), hazard (increased by 1.18%), and vegetation (increased by 2.05%). It is easy to understand why DLCD methods improve performance on the urban land-use change detection. On the one hand, the conventional change detection methods used for water (95.35%), hazard (96.91%), and vegetation (95.18%) applications perform at relatively higher accuracy than the urban land use applications (92.55%); therefore, DLCD methods are more difficult to apply when increasing these relatively high accuracies for water, hazard, and vegetation applications. On the other hand, urban land-use change is caused by human activities, and thus more complicated than the water, hazard, and vegetation changes caused by natural processes. DLCD performs most effectively when addressing complex tasks (Ma et al. Citation2019a).

4.2. Multiclass changes

The overall accuracy of multiclass changes for urban land use, water, hazard, and vegetation applications using the conventional change detection and the DLCD methods is shown in . With regards to these four applications: urban land use, water, vegetation, and hazard change detection, the overall accuracy values of conventional change detection and DLCD methods are from case studies reported in the representative references where both conventional change detection and DLCD methods are used. For the hazard application, there is only one case study.

Figure 9. Distribution of overall accuracies of multi-class changes for urban land use, water, hazard, and vegetation applications using the conventional change detection (CD) and DLCD methods.

Figure 9. Distribution of overall accuracies of multi-class changes for urban land use, water, hazard, and vegetation applications using the conventional change detection (CD) and DLCD methods.

From , in terms of the statistical accuracy, compared with the traditional change detection methods, DLCD methods also increase the accuracy of these four applications in the same way as binary change detection discussed in Section 4.1. This can be evidenced by the fact that the median overall accuracy of these four applications when using the DLCD methods was higher than instances using conventional change detection methods (urban land use from 91.28% to 96.04%, water from 91.17% to 95.61%, hazard from 73.95% to 93.79%, and vegetation from 94.25% to 97.30%). Variability in the accuracy of the DLCD methods was lower than the conventional change detection methods for these four applications. This can be evidenced by the respective IQR within each application, i.e. IQR using the DLCD methods is smaller than that using the conventional change detection methods. This is because DLCD has more powerful feature representation capabilities.

In addition, the increase in the median overall accuracy for hazard (increase by nearly 20%) is more obvious than that of urban land use (increase by 4.76%), water (increase by 4.45%), and vegetation (increase by 3.05%) applications. However, this conclusion is not very reliable because the research on hazard change detection is sparse with only one case study. We need to do more research in the future to reach a definitive conclusion.

4.3. From-To changes

The overall accuracy of from-to changes for urban land use, water, and vegetation applications using the conventional change detection and the DLCD methods is shown in . With regards to three applications: urban land use, water, and vegetation, the overall accuracy values of conventional change detection and DLCD methods are from case studies in the respective references where both conventional change detection and DLCD methods are used. For water and vegetation change detection applications, there is only one case study. For hazard detection, there are no case studies.

Figure 10. Distribution of overall accuracies of from-to changes for urban land use, water, hazard, and vegetation applications using the conventional change detection (CD) and DLCD methods.

Figure 10. Distribution of overall accuracies of from-to changes for urban land use, water, hazard, and vegetation applications using the conventional change detection (CD) and DLCD methods.

From , DLCD increases the accuracy of these three applications in contrast to traditional change detection methods. This can be evidenced by the fact that the median the overall accuracy of these three applications using the DLCD methods was far greater than the accuracy using the conventional change detection methods (urban land use from 85.22% to 95.68%, water from 85.14% to 98.42%, and vegetation from 71.75% to 91.65%). The second conclusion is that the variability of the accuracy of the DLCD methods was lower than the conventional change detection methods for these three applications. This can be evidenced by the respective IQR within each application, i.e. IQR using the DLCD methods is smaller than that using the conventional change detection methods. This is because of the feature representation capabilities for from-to change detection in the DLCD methods.

In addition, the increase in the median overall accuracy for vegetation (increased by 19.90%) and water (increased by 13.28%) change detection is more apparent than the overall accuracy of urban land-use change detection applications (increased by 10.46%) when using DLCD methods. The most likely cause may be that the urban land use from-to change detection using the DLCD methods has more case studies as compared with vegetation and water change detection applications, decreasing the median overall accuracy. From-to change detection case studies however are too few, requiring additional research to reach a more credible conclusion.

According to the existing literature, from-to change detection accuracy improvement using DLCD increased by nearly 10–20%. Next comes multi-class change detection which increased by 3%–20%. The conventional detection methods for binary changes were more than 92.5% and performed at a relatively higher median accuracy than the multi-class change detection from 73.95% to 94.25%, and from-to change detection increased from 71.75% to 85.22% in four applications. Therefore, DLCD methods do not increase these relatively high accuracies in binary change detection applications. From-to changes involve many classification types, and are more complicated than binary and multi-class changes. In the existing literature, it has been shown that DLCD performs optimally in complex tasks (Ma et al. Citation2019a). However, there is limited published work on the multi-class and from-to change detection applications using DLCD. A larger sample of case studies is needed to confirm this conclusion.

5. Dilemmas of DLCD

Although it is evident that the introduction of deep learning improves information representation, change detection methods, and performance in change detection, the limitations on DLCD are not yet clear. The reference studies have indicated two major dilemmas are often encountered when deep learning is applied in a specific application, large sets of labeled training samples are required, and DLCD places high demands on hardware and software. This is also true in change detection. Therefore, in the following subsections, we will examine these dilemmas related to DLCD.

5.1. Training sample dilemma

The training sample dilemma adds complications when deep learning is incorporated into change detection. Deep learning can execute change detection tasks when there are relatively abundant labeled training samples (LeCun, Bengio, and Hinton Citation2015; Gong et al. Citation2019b; Wang et al. Citation2018a). However, the required sample size is well beyond what is available for change detection (Ball, Anderson, and Chan Citation2018; Reichstein et al. Citation2019; De et al. Citation2017). This poses a big challenge for DLCD, for which no research is available. The lack of such research may become a major hurdle for the further application of deep learning in change detection. Therefore, in this section, we summarized the various solutions that have been proposed to this dilemma.

Except for the manual method, we grouped the remaining solutions into two categories: 1) generating large training samples, and 2) adapting to small training samples. The definition, advantages, limitations, and examples of each solution are shown in . The first published publication and the most cited publication for each solution are listed as examples.

Table 3. The definition, advantages, limitations, and examples of solutions to the training sample dilemma in DLCD.

From , the first category is to enhance the size of training samples. The second category is to develop DLCD methods to enhance the efficiency of the training samples. The review then continues with a more detailed discussion of each solution.

5.1.1. Generating large training samples

In this sub-section, we divide the methods for generating large training samples into three sub-categories: Data Augmentation Method (DAM); Supervised Change Detection Method (SCDM); and Unsupervised Change Detection Method (USCDM). DAM is a method of utilizing affine transformations to enhance the size and diversity of training samples. It includes one step. For this method, four basic transformation operations including rotation (Zhan et al. Citation2017; Nemoto et al. Citation2017; Wang et al. Citation2018a; Zhu et al. Citation2018), flip (Zhan et al. Citation2017), mirror (Zhu et al. Citation2018), and cropping (Nemoto et al. Citation2017; Zhan et al. Citation2017) are directly used to generate augmented training samples. The advantage of this method is that it is easy to implement and does not change the spectral or topological information for the training samples. Data augmentation (i.e. flips, mirror, rotations, and cropping) for training samples diversifies its holistic spatial layout and orientation (Yu et al. Citation2017), which can effectively avoid the over-fitting problem in deep learning (Cireşan et al. Citation2010; Luus et al. Citation2015; Simard, Steinkraus, and Platt Citation2003). However, the disadvantage is that it needs an initial training set and the accuracy of the extended training samples directly depends on initial training samples. Recently, beyond these basic transformation operations (i.e. flips, mirror, rotations, and cropping), Gong et al. (Citation2019b) utilized GANs to generate the new training samples given its advantages of generating new information. To extend accurate training samples, SCDM was developed.

SCDM refers to using traditional SCDM such as the supervised object-based change detection method to extend training samples (Zhang et al. Citation2016b). It includes two steps. The traditional SCDM is used to obtain the initial change detection result and the training samples are selected from this initial change detection result. The advantage of this method is that it can generate more accurate training samples than that using the traditional USCDM. However, the disadvantage is that two typical techniques (i.e. the supervised change detection and the subsequent supervised DLCD) are combined, which makes the structure of the whole model more complex. USCDM was developed to eliminate the need for initial training samples.

USCDM refers to using traditional USCDMs to generate change and no change training samples. It includes two steps. The traditional USCDMs are used to obtain the initial change detection result and the training samples are selected from this initial change detection result. Unsupervised pixel-based change detection methods include thresholding methods (Liu et al. Citation2016b), level set methods (Liu et al. Citation2016b), CVA (Zhang, and Zhang Citation2016), and clustering methods (Geng et al. Citation2017) are used to generate training samples. Unsupervised object-based methods including an ensemble learning method based on objects (Gong et al. Citation2017b) and an unsupervised object-based Markov random field (Li, Xu, and Liu Citation2018) are also utilized to generate training samples. Compared with pixel-based methods, the object-based methods can include neighboring information and edge information to generate more accurate training samples. Compared with DAM and SCDM, the advantage of USCDM is that it does not need an initial training set. However, the disadvantage is that the generated training samples are less accurate. In addition to the DLCD methods that generate large training samples, DLCD methods that adapt to small training samples have also been developed.

5.1.2. Adapting to small training samples

We divided the methods that adapt to small training samples into three sub-categories: Semi-Supervised Methods (SSMs), Transfer Learning Method (TLM), and Unsupervised DLCD Methods (UDLCDMs). SSM refers to methods that use a combination of labeled and unlabeled training samples for change detection (Connors, and Vatsavai Citation2017; Gong et al. Citation2019b). It includes two steps. Unlabeled training samples are used to train an unsupervised deep learning network to extract relevant feature information, and this feature information is input into the supervised classifier for change detection using the labeled training samples. The advantage of the SSM is that it only needs a few training samples, which can reduce the expense of manually obtaining training samples. The disadvantage is that it cannot be transferred to other multi-date images; TLM overcomes this shortcoming.

TLM is a method that uses a pre-trained model from other data to detect changes in current multi-date images. In the DLCD literature, there are two popular ways to transfer learning, direct application and fine-tuning. Direct application means using pre-trained neural networks from other kinds of data to extract deep features (Saha, Bovolo, and Bruzzone Citation2019; El Amin, Liu, and Wang Citation2017). Fine-tuning means using small training samples from the current multi-date images to fine-tune a pre-trained model that came from other training datasets for change detection (Waldeland, Reksten, and Salberg Citation2018). Fine-tuning yields more accurate change detection results than direct application. The advantage of TLM is that it only needs a few training samples, and this model can be transferred to new multi-date images. However, the disadvantage is that the transferability depends on the spectral similarity between the training data and the target image (Yosinski et al. Citation2014). UDLCDM overcomes the dependency on spectral similarity and initial training sample requirements.

UDLCDM refers to methods that combine unsupervised DLNNs with USCDMs for change detection. It includes three steps. Unsupervised DLNNs such as DBN (Zhang et al. Citation2016a) and AE (Su et al. Citation2016; Liu et al. Citation2016a) are feature learning tools that are used to effectively extract high-dimension features, and USCDMs such as CVA (Zhang et al. Citation2016a; Su et al. Citation2016) is used to map these features to characterize change information. Clustering methods are used to detect changes. As compared with other methods, the advantage of UDLCDM is that it is automatic and does not need training samples. However, the disadvantage is that the application scenarios are constrained by the limitations of unsupervised DLNNs and unsupervised change detection.

5.2. Hardware and software dilemmas

Compared with the conventional change detection methods, DLCD methods have stricter software and hardware requirements. Considering hardware requirements, conventional change detection methods just need a CPU card, but DLCD often requires a computer with a GPU card. In terms of software requirements, many conventional change detection methods are ready-to-use tools in software such as ENVI, ERDAS, and ArcGIS. DLCD however, not only requires open-source deep learning frameworks such as Caffe/Caffe2.0, Pytorch, Theano, Tensorflow, Keras, and MATLAB but also requires custom programming. A detailed discussion of popular deep learning frameworks can be found in (De Felice Citation2017). Fortunately, the difficulty of applying DLCD is decreasing. For example, ENVI has released the ENVI Deep Learning Module.

6. Future prospects of DLCD

In sections 2, 3, 4, and 5, we compared the difference in how spectral, spatial, temporal, and multi-sensor information were represented between DLCD and conventional change detection methods. We introduced a taxonomy of DLCD methods and provided a systematic comparison of their advantages, limitations, applicability, and performance. We reviewed the difference in the overall improvement in accuracy between binary, multi-class, and from-to change detection for different applications. We reviewed two major limitations in DLCD: training sample and hardware and software dilemmas. In this section, we summarize four future directions: 1) DLCD methods, 2) DLCD applications, 3) training samples, and 4) the implication of remote sensing/change detection/deep learning for DLCD.

6.1. DLCD methods

The DLCD community has made significant progress in developing DLCD methods. However, there are still existing fields to be developed. The future direction of the DLCD methods could be the coupled DLNNs. This can be attributed to its lower accuracy variability (Section 3.2) and the ability to solve specific problems (i.e. resistance to noise, multi-source images and multi-spatial resolution images change detection, time-sensitive change detection, and generating the high-quality information). Specifically, in coupled DLNNs, RM and AM can be potential methods for change detection. For RM, on the one hand, currently, RM has the second-highest median accuracy performance in coupled DLNNs (Section 3.2). On the other hand, time-series change detection is the future direction of change detection, and RM can provide temporal features for it. For AM, although it has the lowest median accuracy performance in coupled DLNNs, it may be the future direction. Compared with other basic deep learning networks, which only have a discriminator neural network (i.e. (0, 1) classifier), AM not only includes a discriminator, but also includes a generator (i.e. generating new information by a noise input). The current AM studies also proved that it has the advantage of generating training samples (Gong et al. Citation2019b) and high-quality change information (Gong et al. Citation2017a). However, the current studies are limited, so it remains a subject for more research.

6.2. DLCD applications

The DLCD community has made significant progress for real applications. However, there are still existing fields to be developed. For urban land use applications, in terms of the spatial extent, all studies focused on the local areas (e.g. 14,400 ha (Mou, Bruzzone, and Zhu Citation2018)) within a city. Therefore, it would be interesting to extend study areas to a larger spatial scale (i.e. regional, national, and global scale). In terms of the detected changes, 49 of the 89 case studies have detected binary changes (El Amin, Liu, and Wang Citation2016; Chu, Cao, and Hayat Citation2016). Next comes multiclass change detection (specific change types are unclear), which has 19 studies (Su et al. Citation2016; Zhang et al. Citation2016a). Only 12 of the 89 studies have detected land use/cover from-to changes (Lyu, Lu, and Mou Citation2016; Lyu, and Lu Citation2017). Although building change detection is very popular in traditional urban change detection, only nine of the 89 urban DLCD studies have detected building changes (Argyridis, and Argialas Citation2016; Nemoto et al. Citation2017). The limited studies for land use/land cover from-to and building change detection can be attributed to the fact that the former involves many change types and the latter involves the building target, more complicated, compared to binary and multiclass change detection. It would be interesting to see more studies on these two types of change detection.

For water applications, most water DLCD studies focused on the DLCD algorithm development and used public data sets. In addition, the study areas are limited. Among the 40 water DLCD studies, 18 were along the yellow river (Zhao et al. Citation2014; Gong et al. Citation2019a; Su, and Cao Citation2018; Zhao et al. Citation2016), four in San Francisco (Gong, Yang, and Zhang Citation2017; Gao et al. Citation2017; Zhao et al. Citation2016), four in the Sulzberger Ice Shelf (Gao et al. Citation2019a, Citation2019b), six for the Weihe River, China (Zhang, and Zhang Citation2016; Gong et al. Citation2017b; Lei et al. Citation2019b), five for Hongqi, China (Gong et al. Citation2017a, Citation2017b; Zhao et al. Citation2017), two in Sardinia, Italy (Zhang et al. Citation2016b; Gong et al. Citation2019a), one for Lake Lotus (Zhang et al. Citation2016a). In the future, we need to conduct more experiments to make the accuracy of these water studies sufficient to meet the requirements of real-world applications.

For hazard applications, there are 42 applications of before- and after-hazard change detection. Among these studies, 31 worked on flooding, including 17 using the same public data set in Ottawa (Zhao et al. Citation2014; Gong et al. Citation2015)) and six using the same public data set in Bern (Zhao et al. Citation2014; Liu et al. Citation2017)), six along the yellow river (Ma et al. Citation2019b; Chen et al. Citation2019), and two in Thailand (Amit, and Aoki Citation2017). Six worked on the landslide, including two in Japan (Amit, and Aoki Citation2017) and four in China (Chen et al. Citation2018; Lei et al. Citation2019a). Two studies mapped the damage caused by the Tohoku tsunami (Sublime, and Kalinicheva Citation2019). One detected the changes before and after the Aere typhoon (Li, Yuan, and Wang Citation2019). One detected the damage by a forest fire (Cao et al. Citation2017). One conducted avalanche detection (Waldeland, Reksten, and Salberg Citation2018). However, there are no damaged buildings, civil war, volcanic eruptions, and droughts change detection studies. It would be interesting to see DLCD used in more hazard scenarios.

Compared with urban land use, water, and hazard applications, the development of vegetation DLCD application lags. To date, only 13 publications have worked on vegetation DLCD in 22 different areas. Among these 22 studies, 20 worked on farmland (Wang et al. Citation2019; Li, Yuan, and Wang Citation2019; Yuan, Wang, and Li Citation2018) and the other two on forests (Khan et al. Citation2017). All farmland studies focused on small areas (e.g. 5670 ha (Yuan, Wang, and Li Citation2018)). Forest studies are too few to analyze their pattern. It would be interesting to see vegetation DLCD in more application scenarios such as mangrove change detection and larger scale (e.g. global scale) in the future.

6.3. Training samples

The DLCD community has made significant progress on the training sample dilemma, with six solutions currently available. These solutions, however, created new problems so it remains a subject for more research. DLCD methods that adapt to small training samples may be a potential area to explore, because this category only needs a few or no training samples, which reduces the manual workload. When generating large training samples, USCDM generated inaccurate training samples, so it remains a subject for more research. In adapting to small training samples, for UDLCDM, the application scenarios were limited by the unsupervised DLNNs and change detection methods, so it remains a subject for more research. In the future, we can make use of the flexible and deep network structure of deep learning to extend the application range of UDLCDM. TLM also has the great potential to solve the training sample problem because it only needs a few training samples, but its transferability is influenced by the distance of the spectral distribution between the training data and the target image, so it also remains a subject for more research. To make the spectral distribution of the training data closer to the target image, the change detection model needs to be trained through repetition and variation such as in a never-ending learning model (Mitchell et al. Citation2018).

6.4. The implication of remote sensing/change detection/deep learning for DLCD

In this section, we discuss the implication of remote sensing, change detection, and deep learning for DLCD. The relationship between DLCD, remote sensing, change detection, and deep learning is shown in .

Figure 11. The relationship between DLCD, remote sensing, change detection, and deep learning.

Figure 11. The relationship between DLCD, remote sensing, change detection, and deep learning.

From , DLCD is the combination of remote sensing, change detection, and deep learning. Thus, we argue that the developments in remote sensing, change detection, and deep learning could be promising directions for developments in future DLCD methods.

The incorporation of spatiotemporal information will play a crucial role in future DLCD developments. The latest work from Yuan et al. (Citation2020) has indicated that spatiotemporal information is indispensable when deep learning is applied to remote sensing applications. With regards to DLCD, a few attempts have been made to exploit spatial and temporal information. For example, Lyu, Lu, and Mou (Citation2016) used the RNN to learn the temporal features. Liu et al. (Citation2017) employed a full-connection layer and a softmax layer to concatenate the output features of two paralleled CNN channels, which amplifies the spatial information. However, spatial and temporal information are considered separately. Recently, one exception is the work of Mou and Zhu (Citation2018) who managed to extract joint spatial-temporal features for land use change detection in complex urban areas using Landsat images by combining CNN and RNN. This study is a good example of simultaneously considering spatial and temporal information by using the same loss function for RNN and CNN. However, this study used shallow CNN and RNN models, which face difficulties in dealing with high-resolution images and long time-series images. In the future, we need to develop a spatiotemporally constrained deep learning model for high-resolution images and long-time-series images.

The incorporation of hybrid methods into DLNNs is anticipated. In the field of change detection, hybrid methods have been used to successfully combine the advantages of separate change detection methods. Currently, in the field of DLCD, hybrid change detection methods have been only used to solve the training samples dilemmas by generating large training samples. Other hybrid strategies such as combining pixel-based and object-based methods have almost been entirely neglected. Therefore, more algorithms that combine DLNNs with hybrid change detection methods are anticipated. In addition, combining different DLNNs may be able to compensate for their single deficiencies, rendering some more reliable results. For example, CNN and SDAE can be combined to take advantage of spatial features from CNN and simultaneously off-load the need for training samples thanks to SDAE (Zhang et al. Citation2016b).

The rapid developments in the deep learning field open new avenues for DLCD as well. Deep learning algorithms such as AlexNet (Han et al. Citation2017b), ResNet (Zhu et al. Citation2021), GoogleNet (Bazi et al. Citation2019), Unet (Jiao et al. Citation2020), Graph Convolutional Network (GCN) (Hong et al. Citation2020a; Gao et al. Citation2021), Spectral Former (Hong et al. Citation2021), and multimodal deep learning framework (Hong et al. 2020), which have been extensively used for remote sensing, are the potential models for DLCD. For example, Hong et al. (2020) have developed a new supervised version of GCNs. This model can jointly use CNNs and GCNs for extracting more diverse and discriminative feature representations for the hyperspectral image classification task, which can potentially model for change detection. To date, a few attempts have been made to exploit these algorithms for DLCD. For example, Waldeland, Reksten, and Salberg (Citation2018) used a ResNet pre-trained from ImageNet data to detect avalanche in SAR images and found that additional training was needed for avalanche detection. Wang et al. (Citation2018a) used a 50-layer Residual Net to obtain differencing feature maps of multi-date high-resolution remote sensing images for change detection. However, this model has only been used to detect simple land cover changes; this study confronts challenges for complex urban land use change detections. Therefore, an immediate need for future studies is to incorporate AlexNet, ResNet, GoogleNet, Unet, GCN, Spectral Former, and multimodal deep learning framework algorithms in more DLCD applications.

7. Conclusion

Change detection permits more effective management and monitoring of natural resources and environmental change. However, as remote sensing images form big data, finding the real change in the multi-date images is very difficult. The emergence of deep learning provides an opportunity for change detection. In this paper, we review DLCD literature to reveal the theoretical underpinnings in five ways: improved information representations, improved change detection methods, performance enhancements, dilemmas of DLCD, and prospects of DLCD. Compared to conventional change detection, DLCD brings advantages in information representation and change detection methods which result in refined performance. Nevertheless, DLCD still faces challenges in lacking training samples and requiring more advanced hardware and software. Finally, we envision the future research of DLCD to improve DLCD methods from the perspective of coupled DLNNs; to widen DLCD applications in land-use/land cover from-to, building, civil war, volcanic eruption, and droughts change detection; to deal with small training samples from the perspective of UDLCDM and TLM; and to absorb developments in remote sensing, change detection, and deep learning to the field of DLCD. We hope this review makes it easier for researchers to find DLCD methods that are most appropriate to their specific applications, to understand the benefits and shortcomings of DLCD, and to contribute to the future development of DLCD .

Acknowledgments

The authors are grateful to anonymous reviewers whose constructive and valuable comments greatly helped us to improve the paper.

Disclosure statement

No potential conflict of interest was reported by the author(s).

Data availability statement (DAS)

The data that support the findings of this study are available from the corresponding author [L. Wang], upon reasonable request.

Additional information

Notes on contributors

Ting Bai

Ting Bai received her BS degree in geographic information systems from Huazhong Agricultural University, Wuhan, China, in 2014 and her PhD degree in photogrammetry and remote sensing from Wuhan University, Wuhan, China, in 2021. She is an engineer in Wuhan Land Arranging Storage Center. Her current research interests include remote sensing and feature fusion, machine learning, ensemble learning, deep learning, and land use and land cover change detection.

Le Wang

Le Wang received his BS degree in Wuhan Technical University of Surveying and Mapping, Wuhan, China, in 1996. He got his MS degree in remote Sensing from Peking University, Beijing, China in 1996. He got his PhD degree in environmental science from University of California, Berkeley in 2003. He is a professor in the State University of New York at Buffalo. His current research interests include remote sensing; geoscience; forest characterization; environment modeling; land cover and land-use change; urban population estimation; invasive species modeling; spatio-temporal analysis and modeling.

Dameng Yin

Dameng Yin received her BS degree in physics in 2010 from Beijing Normal University, China, her ME degree in remote sensing in 2013 from Beijing Normal University, China, and her PhD degree in 2021 from University at Buffalo, USA. She is an assistant professor in the Institute of Crop Sciences, Chinese Academy of Agricultural Sciences, Beijing, China. Her current research focuses on the use of UAV remote sensing and LiDAR for crop phenotyping. Dr. Yin is a member of the American Association of Geographers (AAG) and the International Association of Chinese Professionals in Geographic Information Sciences (CPGIS). She was a recipient of the Third-Place Award of the AAG Remote Sensing Specialty Group (RSSG) Student Honors Paper Competition in 2017, the Second-Place Award of the CPGIS Best Student Paper Competition in 2017, and the Sigma Xi ”Companions in Zealous Research” Award in the University at Buffalo Chapter in 2020.

Kaimin Sun

Kaimin Sun received his BS, MS, and PhD degrees in photogrammetry and remote sensing from Wuhan University, Wuhan, China, in 1999, 2004, and 2008, respectively. He is a professor in the State Key Laboratory of Information Engineering in Surveying, Mapping, and Remote Sensing, Wuhan University. His research interests include photogrammetry, object-oriented image analysis, and image change detection.

Yepei Chen

Yepei Chen received her PhD degree in photogrammetry and remote sensing from Wuhan University, Wuhan, China, in 2021. She is a lecturer in School of Computer Science, Hubei University of Technology.Her current research interests include time series analysis and change detection.

Wenzhuo Li

Wenzhuo Li received his BS and PhD degrees in photogrammetry and remote sensing from Wuhan University, Wuhan, China, in 2011 and 2017, respectively. He is a postdoctoral researcher in State Key Laboratory of Information Engineering in Surveying, Mapping and Remote Sensing, Wuhan University. His current research interests include image segmentation, image classification, land use, and land cover changes detection, and object-oriented image analysis.

Deren Li

Deren Li received the PhD degree in photogrammetry and remote sensing from the University of Stuttgart, Stuttgart, Germany, in 1984. He is an Academician of the Chinese Academy of Sciences, the Chinese Academy of Engineering, and the Euro-Asia International Academy of Sciences. His research interests are spatial information science and technology represented by RS, GPS, and GIS.

References

  • Amit, S.N.K.B., and Y. Aoki. 2017. “Disaster Detection from Aerial Imagery with Convolutional Neural Network.” 2017 International Electronics Symposium on Knowledge Creation and Intelligent Computing (IES-KCIC), 239–245. doi:10.1109/KCIC.2017.8228593.
  • Arabi, M.E.A., M.S. Karoui, and K. Djerriri. 2018. “Optical Remote Sensing Change Detection Through Deep Siamese Network.” In IGARSS 2018 - 2018 Ieee International Geoscience and Remote Sensing Symposium, 5041–5044. doi:10.1109/IGARSS.2018.8518178.
  • Argyridis, A., and D.P. Argialas. 2016. “Building Change Detection Through Multi-Scale GEOBIA Approach by Integrating Deep Belief Networks with Fuzzy Ontologies.” International Journal of Image and Data Fusion 7 (2): 148–171. doi:10.1080/19479832.2016.1158211.
  • Ball, J.E., D.T. Anderson, and C.S. Chan. 2018. “Special Section Guest Editorial: Feature and Deep Learning in Remote Sensing Applications.” Journal of Applied Remote Sensing 11 (4): 042601. doi:10.1117/1.JRS.11.042601.
  • Bazi, Y., M.M. Rahhal, H. Alhichri, and N. Alajlan. 2019. “Simple Yet Effective Fine-Tuning of Deep Cnns Using an Auxiliary Classification Loss for Remote Sensing Scene Classification.” Remote Sensing 11 (24): 2908. doi:10.3390/rs11242908.
  • Bengio, Y., P. Simard, and P. Frasconi. 1994. “Learning Long-Term Dependencies with Gradient Descent is Difficult.” IEEE Transactions on Neural Networks 5 (2): 157–166. doi:10.1109/72.279181.
  • Bovolo, F., L. Bruzzone, and M. Marconcini. 2008. “A Novel Approach to Unsupervised Change Detection Based on a Semisupervised SVM and a Similarity Measure.” IEEE Transactions on Geoscience and Remote Sensing 46 (7): 2070–2082. doi:10.1109/TGRS.2008.916643.
  • Cao, G., Y. Li, Y. Liu, and Y. Shang. 2014. “Automatic Change Detection in High-Resolution Remote-Sensing Images by Means of Level Set Evolution and Support Vector Machine Classification.” International Journal of Remote Sensing 35 (16): 6255–6270. doi:10.1080/01431161.2014.951740.
  • Cao, G., B. Wang, H.C. Xavier, D. Yang, and J. Southworth. 2017. “A New Difference Image Creation Method Based on Deep Neural Networks for Change Detection in Remote-Sensing Images.” International Journal of Remote Sensing 38 (23): 7161–7175. doi:10.1080/01431161.2017.1371861.
  • Cao, C., S. Dragićević, and S. Li. 2019. “Land-Use Change Detection with Convolutional Neural Network Methods.” Environments 6 (2): 25. doi:10.3390/environments6020025.
  • Chan, J.C.W., K.P. Chan, and A.G.O. Yeh. 2001. “Detecting the Nature of Change in an Urban Environment: A Comparison of Machine Learning Algorithms.” Photogrammetric Engineering and Remote Sensing 67 (2): 213–226. doi:10.1144/petgeo.7.1.81.
  • Chehata, N., C. Orny, S. Boukir, D. Guyon, and J.P. Wigneron. 2014. “Object-Based Change Detection in Wind Storm-Damaged Forest Using High-Resolution Multispectral Images.” International Journal of Remote Sensing 35 (13): 4758–4777. doi:10.1080/01431161.2014.930199.
  • Chen, J., P. Gong, C. He, R. Pu, and P. Shi. 2003. “Land-Use/land-Cover Change Detection Using Improved Change-Vector Analysis.” Photogrammetric Engineering and Remote Sensing 69 (4): 369–379. doi:10.14358/PERS.69.4.369.Chen.
  • Chen, F., J. Shi, and M. Gong. 2016. “Differencing Neural Network for Change Detection in Synthetic Aperture Radar Images.” International Conference on Bio-Inspired Computing: Theories and Applications 431–437. doi:10.1007/978-981-10-3611-8_38.Chen.
  • Chen, Z., Y. Zhang, C. Ouyang, F. Zhang, and J. Ma 2018. “Automated Landslides Detection for Mountain Cities Using Multi-Temporal Remote Sensing Imagery.” Sensors 18 (3): 821. doi:10.3390/s18030821.
  • Chen, H., L. Jiao, M. Liang, F. Liu, S. Yang, and B. Hou 2019. “Fast Unsupervised Deep Fusion Network for Change Detection of Multitemporal SAR Images.” Neurocomputing 332: 56–70. doi:10.1016/j.neucom.2018.11.077.
  • Chen, L., F. Rottensteiner, and C. Heipke. 2021. “Feature Detection and Description for Image Matching: From Hand-Crafted Design to Deep Learning.” Geo-Spatial Information Science 24 (1): 58–74. doi:10.1080/10095020.2020.1843376.
  • Cheng, G., X. Xie, J. Han, L. Guo, and G.S. Xia. 2020. “Remote Sensing Image Scene Classification Meets Deep Learning: Challenges, Methods, Benchmarks, and Opportunities.” IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing 13: 3735–3756. doi:10.1109/JSTARS.2020.3005403.
  • Chu, Y., G. Cao, and H. Hayat. 2016. “Change Detection of Remote Sensing Image Based on Deep Neural Networks.” Proceedings of the 2016 2nd International Conference on Artificial Intelligence and Industrial Engineering 133 (1): 262–267. doi:10.2991/aiie-16.2016.61.
  • Cireşan, D.C., U. Meier, L.M. Gambardella, and J. Schmidhuber. 2010. “Deep, Big, Simple Neural Nets for Handwritten Digit Recognition.” Neural Computation 22 (12): 3207–3220. doi:10.1162/NECO_a_00052.
  • Connors, C., and R.R. Vatsavai. 2017. “Semi-Supervised Deep Generative Models for Change Detection in Very High Resolution Imagery.” In 2017 IEEE International Geoscience and Remote Sensing Symposium (IGARSS), 1063–1066. doi:10.1109/IGARSS.2017.8127139.
  • Dai, X.L., and S. Khorram. 1999. “Remotely Sensed Change Detection Based on Artificial Neural Networks.” Photogrammetric Engineering and Remote Sensing 65 (10): 1187–1194. doi:10.1117/1.482717.
  • De, S., D. Pirrone, F. Bovolo, L. Bruzzone, and A. Bhattacharya. 2017. “A Novel Change Detection Framework Based on Deep Learning for the Analysis of Multi-Temporal Polarimetric Sar Images.” In 2017 IEEE International Geoscience and Remote Sensing Symposium (IGARSS), 5193–5196. doi:10.1109/IGARSS.2017.8128171.
  • De Felice, M. 2017. “Which Deep Learning Network is Best for You? IDG Communications.” accessed May 4 20 1 7. http://www.cio.com/article/3193689/artificial-intelligence/which-deep-learning-network-is-best-for-you.html
  • El Amin, A.M., Q. Liu, and Y. Wang. 2016. “Convolutional Neural Network Features Based Change Detection in Satellite Images.” First International Workshop on Pattern Recognition, no. 10011: 181–186. doi:10.1117/12.2243798.
  • El Amin, A.M., Q. Liu, and Y. Wang. 2017. “Zoom Out Cnns Features for Optical Remote Sensing Change Detection.” In 2017 2nd International Conference on Image, Vision and Computing (ICIVC), 812–817. doi:10.1109/ICIVC.2017.7984667.
  • Gao, F., X. Liu, J. Dong, G. Zhong, and M. Jian. 2017. “Change Detection in SAR Images Based on Deep Semi-NMF and SVD Networks.” Remote Sensing 9 (5): 435. doi:10.3390/rs9050435.
  • Gao, F., X. Wang, Y. Gao, J. Dong, and S. Wang. 2019a. “Sea Ice Change Detection in SAR Images Based on Convolutional-Wavelet Neural Networks.” IEEE Geoscience and Remote Sensing Letters 16 (8): 1240–1244. doi:10.1109/LGRS.2019.2895656.
  • Gao, Y., F. Gao, J. Dong, and S. Wang. 2019b. “Transferred Deep Learning for Sea Ice Change Detection from Synthetic-Aperture Radar Images.” IEEE Geoscience Remote Sensing Letters 16 (10): 1655–1659. doi:10.1109/LGRS.2019.2906279.
  • Gao, Y., J. Shi, J. Li, and R. Wang. 2021. “Remote Sensing Scene Classification Based on High-Order Graph Convolutional Network.” European Journal of Remote Sensing 54 (sup1): 141–155. doi:10.1080/22797254.2020.1868273.
  • Geng, J., H. Wang, J. Fan, and X. Ma. 2017. “Change Detection of SAR Images Based on Supervised Contractive Autoencoders and Fuzzy Clustering.” In 2017 International Workshop on Remote Sensing with Intelligent Processing (RSIP), 1–3. doi:10.1109/RSIP.2017.7958819.
  • Gong, P. 1993. “Change Detection Using Principal Component Analysis and Fuzzy Set Theory.” Canadian Journal of Remote Sensing 19 (1): 22–29. doi:10.1080/07038992.1993.10855147.
  • Gong, M., J. Zhao, J. Liu, Q. Miao, and L. Jiao. 2015. “Change Detection in Synthetic Aperture Radar Images Based on Deep Neural Networks.” IEEE Transactions on Neural Networks and Learning Systems 27 (1): 125–138. doi:10.1109/TNNLS.2015.2435783.
  • Gong, M., H. Yang, and P. Zhang. 2017. “Feature Learning and Change Feature Classification Based on Deep Learning for Ternary Change Detection in SAR Images.” ISPRS Journal of Photogrammetry and Remote Sensing 129: 212–225. doi:10.1016/j.isprsjprs.2017.05.001.
  • Gong, M., X. Niu, P. Zhang, and Z. Li. 2017a. “Generative Adversarial Networks for Change Detection in Multispectral Imagery.” IEEE Geoscience and Remote Sensing Letters 14 (12): 2310–2314. doi:10.1109/Lgrs.2017.2762694.
  • Gong, M., T. Zhan, P. Zhang, and Q. Miao. 2017b. “Superpixel-Based Difference Representation Learning for Change Detection in Multispectral Remote Sensing Images.” IEEE Transactions on Geoscience and Remote Sensing 55 (5): 2658–2673. doi:10.1109/TGRS.2017.2650198.
  • Gong, M., X. Niu, T. Zhan, and M. Zhang. 2019a. “A Coupling Translation Network for Change Detection in Heterogeneous Images.” International Journal of Remote Sensing 40 (9): 3647–3672. doi:10.1080/01431161.2018.1547934.
  • Gong, M., Y. Yang, T. Zhan, X. Niu, and S. Li. 2019b. “A Generative Discriminatory Classified Network for Change Detection in Multispectral Imagery.” IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing 12 (1): 321–333. doi:10.1109/JSTARS.2018.2887108.
  • Goodfellow, I., J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair, A. Courville, and Y. Bengio. 2014. “Generative Adversarial Nets.” Advances in Neural Information Processing Systems 27. doi:10.5555/2969033.2969125.
  • Han, J., X. Meng, X. Zhou, B. Yi, M. Liu, and W.N. Xiang. 2017a. “A Long-Term Analysis of Urbanization Process, Landscape Change, and Carbon Sources and Sinks: A Case Study in China’s Yangtze River Delta Region.” Journal of Cleaner Production 141: 1040–1050. doi:10.1016/j.jclepro.2016.09.177.
  • Han, X., Y. Zhong, L. Cao, and L. Zhang. 2017b. “Pre-Trained Alexnet Architecture with Pyramid Pooling and Supervision for High Spatial Resolution Remote Sensing Image Scene Classification.” Remote Sensing 9 (8): 848. doi:10.3390/rs9080848.
  • Healey, S.P., W.B. Cohen, Z. Yang, C.K. Brewer, E.B. Brooks, N. Gorelick, A.J. Hernandez, et al. 2018. ”Mapping Forest Change Using Stacked Generalization: An Ensemble Approach”. Remote Sensing of Environment 204: 717–728. 10.1016/j.rse.2017.09.029.
  • Hinton, G.E., S. Osindero, and Y.W. Teh. 2006. “A Fast Learning Algorithm for Deep Belief Nets.” Neural Computation 18 (7): 1527–1554. doi:10.1162/neco.2006.18.7.1527.
  • Hinton, G.E., and R.R. Salakhutdinov. 2006. “Reducing the Dimensionality of Data with Neural Networks.” Science 313 (5786): 504–507. doi:10.1126/science.1127647.
  • Hong, D., N. Yokoya, J. Chanussot, and X.X. Zhu. 2018. “An Augmented Linear Mixing Model to Address Spectral Variability for Hyperspectral Unmixing.” IEEE Transactions on Image Processing 28 (4): 1923–1938. doi:10.1109/TIP.2018.2878958.
  • Hong, D., L. Gao, J. Yao, B. Zhang, A. Plaza, and J. Chanussot. 2020a. “Graph Convolutional Networks for Hyperspectral Image Classification.” IEEE Transactions on Geoscience and Remote Sensing 59 (7): 5966–5978. doi:10.1109/TGRS.2020.3015157.
  • Hong, D., L. Gao, N. Yokoya, J. Yao, J. Chanussot, Q. Du, and B. Zhang. 2020b. “More Diverse Means Better: Multimodal Deep Learning Meets Remote-Sensing Imagery Classification.” IEEE Transactions on Geoscience and Remote Sensing 59 (5): 4340–4354. doi:10.1109/TGRS.2020.3016820.
  • Hong, D., Z. Han, J. Yao, L. Gao, B. Zhang, A. Plaza, and J. Chanussot. 2021. “SpectralFormer: Rethinking Hyperspectral Image Classification with Transformers.” IEEE Transactions on Geoscience and Remote Sensing 60: 1–15. doi:10.1109/TGRS.2021.3130716.
  • Howarth, P.J., and G.M. Wickware. 1981. “Procedures for Change Detection Using Landsat Digital Data.” International Journal of Remote Sensing 2 (3): 277–291. doi:10.1080/01431168108948362.
  • Hussain, M., D. Chen, A. Cheng, H. Wei, and D. Stanley. 2013. “Change Detection from Remotely Sensed Images: From Pixel-Based to Object-Based Approaches.” ISPRS Journal of Photogrammetry and Remote Sensing 80: 91–106. doi:10.1016/j.isprsjprs.2013.03.006.
  • Iino, S., R. Ito, K. Doi, T. Imaizumi, and S. Hikosaka. 2018. “CNN-Based Generation of High-Accuracy Urban Distribution Maps Utilising SAR Satellite Imagery for Short-Term Change Monitoring.” International Journal of Image and Data Fusion 9 (4): 302–318. doi:10.1080/19479832.2018.1491897.
  • Jiao, L., L. Huo, C. Hu, and P. Tang. 2020. “Refined Unet: Unet-Based Refinement Network for Cloud and Shadow Precise Segmentation.” Remote Sensing 12 (12): 2001. doi:10.3390/rs12122001.
  • Johnson, R.D., and E.S. Kasischke. 1998. “Change Vector Analysis: A Technique for the Multispectral Monitoring of Land Cover and Condition.” International Journal of Remote Sensing 19 (3): 411–426. doi:10.1080/014311698216062.
  • Kemker, R., C. Salvaggio, and C. Kanan. 2018. “Algorithms for Semantic Segmentation of Multispectral Remote Sensing Imagery Using Deep Learning.” ISPRS Journal of Photogrammetry and Remote Sensing 145: 60–77. doi:10.1016/j.isprsjprs.2018.04.014.
  • Khan, S.H., X. He, F. Porikli, and M. Bennamoun. 2017. “Forest Change Detection in Incomplete Satellite Images with Deep Neural Networks.” IEEE Transactions on Geoscience and Remote Sensing 55 (9): 5407–5423. doi:10.1109/TGRS.2017.2707528.
  • Khelifi, L., and M. Mignotte. 2020. “Deep Learning for Change Detection in Remote Sensing Images: Comprehensive Review and Meta-Analysis.” IEEE Access 8: 126385–126400. doi:10.1109/ACCESS.2020.3008036.
  • LeCun, Y., L. Bottou, Y. Bengio, and P. Haffner. 1998. “Gradient-Based Learning Applied to Document Recognition.” Proceedings of the IEEE 86 (11): 2278–2324. doi:10.1109/5.726791.
  • LeCun, Y., Y. Bengio, and G. Hinton. 2015. “Deep Learning.” Nature 521 (7553): 436–444. doi:10.1038/nature14539.
  • Lei, T., Q. Zhang, D. Xue, T. Chen, H. Meng, and A.K. Nandi. 2019a. “End-To-End Change Detection Using a Symmetric Fully Convolutional Network for Landslide Mapping.” In ICASSP 2019-2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 3027–3031. doi: 10.1109/ICASSP.2019.8682802.
  • Lei, Y., X. Liu, J. Shi, C. Lei, and J. Wang. 2019b. “Multiscale Superpixel Segmentation with Deep Features for Change Detection.” IEEE Access 7: 36600–36616. doi:10.1109/ACCESS.2019.2902613.
  • Li, Y., L. Zhou, G. Lu, B. Hou, and L. Jiao. 2017. “Change Detection in Synthetic Aperture Radar Images Based on Log-Mean Operator and Stacked Auto-Encoder.” In 2017 IEEE International Geoscience and Remote Sensing Symposium (IGARSS), 3090–3306. doi:10.1109/IGARSS.2017.8127652.
  • Li, Y., L. Xu, and T. Liu. 2018. “Unsupervised Change Detection for Remote Sensing Images Based on Object-Based MRF and Stacked Autoencoders.” 2016 International Conference on Orange Technologies (ICOT), 64–67. doi:10.1109/ICOT.2016.8278980.
  • Li, X., Z. Yuan, and Q. Wang. 2019. “Unsupervised Deep Noise Modeling for Hyperspectral Image Change Detection.” Remote Sensing 11 (3): 258. doi:10.3390/rs11030258.
  • Li, Y., C. Peng, Y. Chen, L. Jiao, L. Zhou, and R. Shang. 2019. “A Deep Learning Method for Change Detection in Synthetic Aperture Radar Images.” IEEE Transactions on Geoscience and Remote Sensing 57 (8): 5751–5763. doi:10.1109/Tgrs.2019.2901945.
  • Li, R., S. Zheng, C. Duan, L. Wang, and C. Zhang. 2022. “Land Cover Classification from Remote Sensing Images Based on Multi-Scale Fully Convolutional Network.” Geo-Spatial Information Science 1–17. doi:10.1080/10095020.2021.2017237.
  • Lillesand, T., R.W. Kiefer, and J. Chipman. 2015. Remote Sensing and Image Interpretation. New York, NJ, USA: John Wiley & Sons.
  • Liu, S., L. Bruzzone, F. Bovolo, and P. Du. 2014. “Hierarchical Unsupervised Change Detection in Multitemporal Hyperspectral Images.” IEEE Transactions on Geoscience and Remote Sensing 53 (1): 244–260. doi:10.1109/TGRS.2014.2321277.
  • Liu, J., M. Gong, K. Qin, and P. Zhang. 2016a. “A Deep Convolutional Coupling Network for Change Detection Based on Heterogeneous Optical and Radar Images.” IEEE Transactions on Neural Networks and Learning Systems 29 (3): 545–559. doi:10.1109/TNNLS.2016.2636227.
  • Liu, J., M. Gong, J. Zhao, H. Li, and L. Jiao. 2016b. “Difference Representation Learning Using Stacked Restricted Boltzmann Machines for Change Detection in SAR Images.” Soft Computing 20 (12): 4645–4657. doi:10.1007/s00500-014-1460-0.
  • Liu, T., Y. Li, Y. Cao, and Q. Shen. 2017. “Change Detection in Multitemporal Synthetic Aperture Radar Images Using Dual-Channel Convolutional Neural Network.” Journal of Applied Remote Sensing 11 (4): 042615. doi:10.1117/1.JRS.11.042615.
  • Liu, J., M. Gong, K. Qin, and P. Zhang. 2018. “A Deep Convolutional Coupling Network for Change Detection Based on Heterogeneous Optical and Radar Images.” IEEE Transactions on Neural Networks and Learning Systems 29 (3): 545–559.
  • Liu, T., L. Yang, and D. Lunga. 2021. “Change Detection Using Deep Learning Approach with Object-Based Image Analysis.” Remote Sensing of Environment 256: 112308. doi:10.1016/j.rse.2021.112308.
  • Lu, D., P. Mausel, E. Brondizio, and E. Moran. 2004. “Change Detection Techniques.” International Journal of Remote Sensing 25 (12): 2365–2401. doi:10.1080/0143116031000139863.
  • Luus, F.P., B.P. Salmon, F. Van den Bergh, and B.T.J. Maharaj. 2015. “Multiview Deep Learning for Land-Use Classification.” IEEE Geoscience and Remote Sensing Letters 12 (12): 2448–2452. doi:10.1109/LGRS.2015.2483680.
  • Lyu, H., H. Lu, and L. Mou. 2016. “Learning a Transferable Change Rule from a Recurrent Neural Network for Land Cover Change Detection.” Remote Sensing 8 (6): 506. doi:10.3390/rs8060506.
  • Lyu, H., and H. Lu. 2017. “A Deep Information Based Transfer Learning Method to Detect Annual Urban Dynamics of Beijing and New York from 1984–2016.” In 2017 IEEE International Geoscience and Remote Sensing Symposium (IGARSS), 1958–1961. doi:10.1109/IGARSS.2017.8127363.
  • Ma, L., Y. Liu, X. Zhang, Y. Ye, G. Yin, and B.A. Johnson. 2019a. “Deep Learning in Remote Sensing Applications: A Meta-Analysis and Review.” ISPRS Journal of Photogrammetry and Remote Sensing 152: 166–177. doi:10.1016/j.isprsjprs.2019.04.015.
  • Ma, W., Y. Xiong, Y. Wu, H. Yang, X. Zhang, and L. Jiao. 2019b. “Change Detection in Remote Sensing Images Based on Image Mapping and a Deep Capsule Network.” Remote Sensing 11 (6): 626. doi:10.3390/rs11060626.
  • McDermid, G.J., J. Linke, A.D. Pape, D.N. Laskin, A.J. McLane, and S.E. Franklin. 2008. “Object-Based Approaches to Change Analysis and Thematic Map Update: Challenges and Limitations.” Canadian Journal of Remote Sensing 34 (5): 462–466. doi:10.5589/m08-061.
  • Mercier, G., G. Moser, and S.B. Serpico. 2008. “Conditional Copulas for Change Detection in Heterogeneous Remote Sensing Images.” IEEE Transactions on Geoscience and Remote Sensing 46 (5): 1428–1441. doi:10.1109/TGRS.2008.916476.
  • Mitchell, T., W. Cohen, E. Hruschka, P. Talukdar, B. Yang, J. Betteridge, A. Carlson, et al. 2018. ”Never-Ending Learning.” Communications of the ACM 61 (5): 103–115. DOI:10.1145/3191513.
  • Mou, L., L. Bruzzone, and X. Zhu. 2018. “Learning Spectral-Spatial-Temporal Features via a Recurrent Convolutional Neural Network for Change Detection in Multispectral Imagery.” IEEE Transactions on Geoscience and Remote Sensing 57 (2): 924–935. doi:10.1109/TGRS.2018.2863224.
  • Mou, L., and X. Zhu. 2018. “A Recurrent Convolutional Neural Network for Land Cover Change Detection in Multispectral Images.” In IGARSS 2018 - 2018 IEEE International Geoscience and Remote Sensing Symposium, 4363–4366. doi:10.1109/IGARSS.2018.8517375.
  • Murray, H., A. Lucieer, and R. Williams. 2010. “Texture-Based Classification of Sub-Antarctic Vegetation Communities on Heard Island.” International Journal of Applied Earth Observation and Geoinformation 12 (3): 138–149. doi:10.1016/j.jag.2010.01.006.
  • Nemoto, K., R. Hamaguchi, M. Sato, A. Fujita, T. Imaizumi, and S. Hikosaka. 2017. “Building Change Detection via a Combination of Cnns Using Only RGB Aerial Imageries.” Remote Sensing Technologies and Applications in Urban Environments II 10431: 104310J. doi:10.1117/12.2277912.
  • Newbold, T., L.N. Hudson, S.L. Hill, S. Contu, I. Lysenko, R.A. Senior, L. Börger, et al. 2015. ”Global Effects of Land Use on Local Terrestrial Biodiversity.” Nature 520 (7545): 45–50. doi:10.1038/nature14324.
  • Prendes, J., M. Chabert, F. Pascal, A. Giros, and J.Y. Tourneret. 2015a. “Change Detection for Optical and Radar Images Using a Bayesian Nonparametric Model Coupled with a Markov Random Field.” In 2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 1513–1517. doi:10.1109/ICASSP.2015.7178223.
  • Prendes, J., M. Chabert, F. Pascal, A. Giros, and J.Y. Tourneret. 2015b. “Performance Assessment of a Recent Change Detection Method for Homogeneous and Heterogeneous Images.” Revue Francaise de Photogrammetrie Et de Teledetection 209: 23–29.
  • Radford, A., L. Metz, and S. Chintala. 2015. “Unsupervised Representation Learning with Deep Convolutional Generative Adversarial Networks.” arXiv preprint arXiv:1511.06434. doi:10.48550/arXiv.1511.06434.
  • Reichstein, M., G. Camps-Valls, B. Stevens, M. Jung, J. Denzler, and N. Carvalhais. 2019. “Deep Learning and Process Understanding for Data-Driven Earth System Science.” Nature 566 (7743): 195–204. doi:10.1038/s41586-019-0912-1.
  • Saha, S., F. Bovolo, and L. Bruzzone. 2019. “Unsupervised Deep Change Vector Analysis for Multiple-Change Detection in VHR Images.” IEEE Transactions on Geoscience and Remote Sensing 57 (6): 3677–3693. doi:10.1109/TGRS.2018.2886643.
  • Saito, K., R.J. Spence, C. Going, and M. Markus. 2004. “Using High-Resolution Satellite Images for Post-Earthquake Building Damage Assessment: A Study Following the 26 January 2001 Gujarat Earthquake.” Earthquake Spectra 20 (1): 145–169. doi:10.1193/1.1650865.
  • Shao, Z., and C. Liu. 2014. “The Integrated Use of DMSP-OLS Nighttime Light and MODIS Data for Monitoring Large-Scale Impervious Surface Dynamics: A Case Study in the Yangtze River Delta.” Remote Sensing 6 (10): 9359–9378. doi:10.3390/rs6109359.
  • Shao, Z., and J. Cai. 2018. “Remote Sensing Image Fusion with Deep Convolutional Neural Network.” IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing 11 (5): 1656–1669. doi:10.1109/JSTARS.2018.2805923.
  • Shao, Z., H. Fu, D. Li, O. Altan, and T. Cheng. 2019a. “Remote Sensing Monitoring of Multi-Scale Watersheds Impermeability for Urban Hydrological Evaluation.” Remote Sensing of Environment 232: 111338. doi:10.1016/j.rse.2019.111338.
  • Shao, Z., Y. Pan, C. Diao, and J. Cai. 2019b. “Cloud Detection in Remote Sensing Images Based on Multiscale Features-Convolutional Neural Network.” IEEE Transactions on Geoscience and Remote Sensing 57 (6): 4062–4076. doi:10.1109/TGRS.2018.2889677.
  • Shi, W., M. Zhang, R. Zhang, S. Chen, and Z. Zhan. 2020. “Change Detection Based on Artificial Intelligence: State-Of-The-Art and Challenges.” Remote Sensing 12 (10): 1688. doi:10.3390/rs12101688.
  • Silván-Cárdenas, J.L., and L. Wang. 2014. “On Quantifying Post-Classification Subpixel Landcover Changes.” ISPRS Journal of Photogrammetry and Remote Sensing 98: 94–105. doi:10.1016/j.isprsjprs.2014.09.018.
  • Simard, P.Y., D. Steinkraus, and J.C. Platt. 2003. “Best Practices for Convolutional Neural Networks Applied to Visual Document Analysis.” Seventh International Conference on Document Analysis and Recognition, 958–963. doi:10.1109/ICDAR.2003.1227801.
  • Singh, A. 1989. “Review Article Digital Change Detection Techniques Using Remotely-Sensed Data.” International Journal of Remote Sensing 10 (6): 989–1003. doi:10.1080/01431168908903939.
  • Song, A., J. Choi, Y. Han, and Y. Kim. 2018. “Change Detection in Hyperspectral Images Using Recurrent 3D Fully Convolutional Networks.” Remote Sensing 10 (11): 1827. doi:10.3390/rs10111827.
  • Su, L., J. Shi, P. Zhang, Z. Wang, and M. Gong. 2016. “Detecting Multiple Changes from Multi-Temporal Images by Using Stacked Denosing Autoencoder Based Change Vector Analysis.” In 2016 International Joint Conference on Neural Networks (IJCNN), 1269–1276. doi:10.1109/IJCNN.2016.7727343.
  • Su, L., M. Gong, P. Zhang, M. Zhang, J. Liu, and H. Yang. 2017. “Deep Learning and Mapping Based Ternary Change Detection for Information Unbalanced Images.” Pattern Recognition 66: 213–228. doi:10.1016/j.patcog.2017.01.002.
  • Su, L., and X. Cao. 2018. “Fuzzy Autoencoder for Multiple Change Detection in Remote Sensing Images.” Journal of Applied Remote Sensing 12 (3): 035014. doi:10.1117/1.jrs.12.035014.
  • Sublime, J., and E. Kalinicheva. 2019. “Automatic Post-Disaster Damage Mapping Using Deep-Learning Techniques for Change Detection: Case Study of the Tohoku Tsunami.” Remote Sensing 11 (9): 1123. doi:10.3390/rs11091123.
  • Tang, Y., X. Huang, and L. Zhang. 2013. “Fault-Tolerant Building Change Detection from Urban High-Resolution Remote Sensing Imagery.” IEEE Geoscience and Remote Sensing Letters 10 (5): 1060–1064. doi:10.1109/LGRS.2012.2228626.
  • Tewkesbury, A.P., A.J. Comber, N.J. Tate, A. Lamb, and P.F. Fisher. 2015. “A Critical Synthesis of Remotely Sensed Optical Image Change Detection Techniques.” Remote Sensing of Environment 160: 1–14. doi:10.1016/j.rse.2015.01.006.
  • Volpi, M., G. Camps-Valls, and D. Tuia. 2015. “Spectral Alignment of Multi-Temporal Cross-Sensor Images with Automated Kernel Canonical Correlation Analysis.” ISPRS Journal of Photogrammetry and Remote Sensing 107: 50–63. doi:10.1016/j.isprsjprs.2015.02.005.
  • Waldeland, A.U., J.H. Reksten, and A.B. Salberg. 2018. “Avalanche Detection in Sar Images Using Deep Learning.” In IGARSS 2018-2018 IEEE International Geoscience and Remote Sensing Symposium, 2386–2389. doi:10.1109/IGARSS.2018.8517536.
  • Wang, Q., X. Zhang, G. Chen, F. Dai, Y. Gong, and K. Zhu. 2018a. “Change Detection Based on Faster R-CNN for High-Resolution Remote Sensing Images.” Remote Sensing Letters 9 (10): 923–932. doi:10.1080/2150704x.2018.1492172.
  • Wang, S., D. Quan, X. Liang, M. Ning, Y. Guo, and L. Jiao. 2018b. “A Deep Learning Framework for Remote Sensing Image Registration.” ISPRS Journal of Photogrammetry and Remote Sensing 145: 148–164. doi:10.1016/j.isprsjprs.2017.12.012.
  • Wang, Q., Z. Yuan, Q. Du, and X. Li. 2019. “GETNET: A General End-To-End 2-D CNN Framework for Hyperspectral Image Change Detection.” IEEE Transactions on Geoscience and Remote Sensing 57 (1): 3–13. doi:10.1109/tgrs.2018.2849692.
  • Weng, Q. 2002. “Land Use Change Analysis in the Zhujiang Delta of China Using Satellite Remote Sensing, GIS and Stochastic Modelling.” Journal of Environmental Management 64 (3): 273–284. doi:10.1006/jema.2001.0509.
  • Xiao, R., R. Cui, M. Lin, L. Chen, Y. Ni, and X. Lin. 2018. “SOMDNCD: Image Change Detection Based on Self-Organizing Maps and Deep Neural Networks.” IEEE Access 6: 35915–35925. doi:10.1109/ACCESS.2018.2849110.
  • Xu, Y., S. Xiang, C. Huo, and C. Pan. 2013. “Change Detection Based on Auto-Encoder Model for VHR Images.” Mippr 2013: Pattern Recognition and Computer Vision 8919: 891902. doi:10.1117/12.2031104.
  • Yosinski, J., J. Clune, Y. Bengio, and H. Lipson. 2014. “How Transferable are Features in Deep Neural Networks?” Advances in Neural Information Processing Systems. 3320–3328.
  • Yu, X., X. Wu, C. Luo, and P. Ren. 2017. “Deep Learning in Remote Sensing Scene Classification: A Data Augmentation Enhanced Convolutional Neural Network Framework.” GIScience & Remote Sensing 54 (5): 741–758. doi:10.1080/15481603.2017.1323377.
  • Yuan, F., K.E. Sawaya, B.C. Loeffelholz, and M.E. Bauer. 2005. “Land Cover Classification and Change Analysis of the Twin Cities (Minnesota) Metropolitan Area by Multitemporal Landsat Remote Sensing.” Remote Sensing of Environment 98 (2–3): 317–328. doi:10.1016/j.rse.2005.08.006.
  • Yuan, Z., Q. Wang, and X. Li. 2018. “ROBUST Pcanet for Hyperspectral Image Change Detection.” In IGARSS 2018-2018 IEEE International Geoscience and Remote Sensing Symposium, 4931–4934. doi:10.1109/IGARSS.2018.8518196.
  • Yuan, Q., H. Shen, T. Li, Z. Li, S. Li, Y. Jiang, H. Xu, et al. 2020. ”Deep Learning in Environmental Remote Sensing: Achievements and Challenges”. Remote Sensing of Environment 241: 111716. 10.1016/j.rse.2020.111716.
  • Zhan, Y., K. Fu, M. Yan, X. Sun, H. Wang, and X. Qiu. 2017. “Change Detection Based on Deep Siamese Convolutional Network for Optical Aerial Images.” IEEE Geoscience and Remote Sensing Letters 14 (10): 1845–1849. doi:10.1109/LGRS.2017.2738149.
  • Zhan, T., M. Gong, J. Liu, and P. Zhang. 2018. “Iterative Feature Mapping Network for Detecting Multiple Changes in Multi-Source Remote Sensing Images.” ISPRS Journal of Photogrammetry and Remote Sensing 146: 38–51. doi:10.1016/j.isprsjprs.2018.09.002.
  • Zhang, H., and P. Zhang. 2016. “Deep Difference Representation Learning for Multi-Spectral Imagery Change Detection.” Proceedings of the 2016 5th International Conference on Advanced Materials and Computer Science, 80: 1008–1014. doi:10.2991/icamcs-16.2016.204.
  • Zhang, L., L. Zhang, and B. Du. 2016. “Deep Learning for Remote Sensing Data: A Technical Tutorial on the State of the Art.” IEEE Geoscience and Remote Sensing Magazine 4 (2): 22–40. doi:10.1109/MGRS.2016.2540798.
  • Zhang, H., M. Gong, P. Zhang, L. Su, and J. Shi. 2016a. “Feature-Level Change Detection Using Deep Representation and Feature Change Analysis for Multispectral Imagery.” IEEE Geoscience and Remote Sensing Letters 13 (11): 1666–1670. doi:10.1109/Lgrs.2016.2601930.
  • Zhang, P., M. Gong, L. Su, J. Liu, and Z. Li. 2016b. “Change Detection Based on Deep Feature Representation and Mapping Transformation for Multi-Spatial-Resolution Remote Sensing Images.” ISPRS Journal of Photogrammetry and Remote Sensing 116: 24–41. doi:10.1016/j.isprsjprs.2016.02.013.
  • Zhang, H., X. Ning, Z. Shao, and H. Wang. 2019. “Spatiotemporal Pattern Analysis of China’s Cities Based on High-Resolution Imagery from 2000 to 2015.” ISPRS International Journal of Geo-Information 8 (5): 241. doi:10.3390/ijgi8050241.
  • Zhang, R., Z. Shao, X. Huang, J. Wang, and D. Li. 2020. “Object Detection in UAV Images via Global Density Fused Convolutional Network.” Remote Sensing 12 (19): 3140. doi:10.3390/rs12193140.
  • Zhao, J., M. Gong, J. Liu, and L. Jiao. 2014. “Deep Learning to Classify Difference Image for Image Change Detection.” In 2014 International Joint Conference on Neural Networks (IJCNN), 411–417. doi:10.1109/IJCNN.2014.6889510.
  • Zhao, Q., J. Ma, M. Gong, H. Li, and T. Zhan. 2016. “Three-Class Change Detection in Synthetic Aperture Radar Images Based on Deep Belief Network.” Journal of Computational and Theoretical Nanoscience 13 (6): 3757–3762. doi:10.1166/jctn.2016.5208.
  • Zhao, W., Z. Wang, M. Gong, and J. Liu. 2017. “Discriminative Feature Learning for Unsupervised Change Detection in Heterogeneous Images Based on a Coupled Neural Network.” IEEE Transactions on Geoscience and Remote Sensing 55 (12): 7066–7080. doi:10.1109/TGRS.2017.2739800.
  • Zhou, J., B. Yu, and J. Qin. 2014. “Multi-Level Spatial Analysis for Change Detection of Urban Vegetation at Individual Tree Scale.” Remote Sensing 6 (9): 9086–9103. doi:10.3390/rs6099086.
  • Zhou, L., Z. Shao, S. Wang, and X. Huang. 2022. “Deep Learning-Based Local Climate Zone Classification Using Sentinel-1 SAR and Sentinel-2 Multispectral Imagery.” Geo-Spatial Information Science 1–16. doi:10.1080/10095020.2022.2030654.
  • Zhu, X., D. Tuia, L. Mou, G. Xia, L. Zhang, F. Xu, and F. Fraundorfer. 2017. “Deep Learning in Remote Sensing: A Comprehensive Review and List of Resources.” IEEE Geoscience and Remote Sensing Magazine 5 (4): 8–36. doi:10.1109/MGRS.2017.2762307.
  • Zhu, Z. 2017. “Change Detection Using Landsat Time Series: A Review of Frequencies, Preprocessing, Algorithms, and Applications.” ISPRS Journal of Photogrammetry and Remote Sensing 130: 370–384. doi:10.1016/j.isprsjprs.2017.06.013.
  • Zhu, B., H. Gao, X. Wang, M. Xu, and X. Zhu. 2018. “Change Detection Based on the Combination of Improved SegNet Neural Network and Morphology.” In 2018 IEEE 3rd International Conference on Image, Vision and Computing (ICIVC), 55–59. doi:10.1109/ICIVC.2018.8492747.
  • Zhu, H., M. Ma, W. Ma, L. Jiao, S. Hong, J. Shen, and B. Hou. 2021. “A Spatial-Channel Progressive Fusion ResNet for Remote Sensing Classification.” Information Fusion 70: 72–87. doi:10.1016/j.inffus.2020.12.008.

Appendix

Table A1. Nomenclature.