Next Article in Journal
SI ‘Multiscale Fatigue Design’
Next Article in Special Issue
A Novel in Situ Stress Monitoring Technique for Fracture Rock Mass and Its Application in Deep Coal Mines
Previous Article in Journal
Directional Force Feedback: Mechanical Force Concentration for Immersive Experience in Virtual Reality
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Comprehensive and Automated Fusion Method: The Enhanced Flexible Spatiotemporal DAta Fusion Model for Monitoring Dynamic Changes of Land Surface

1
College of Urban and Environmental Science, Northwest University, Xi’an 710127, China
2
Shaanxi Key Laboratory of Earth Surface System and Environmental Carrying Capacity, Northwest University, Xi’an 710127, China
3
Shaanxi Key Laboratory of Ecology and Environment of River Wetland, Weinan Normal University, Weinan 714099, China
4
College of Resources and Environment, University of Chinese Academy of Sciences, Beijing 100049, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2019, 9(18), 3693; https://doi.org/10.3390/app9183693
Submission received: 30 July 2019 / Revised: 3 September 2019 / Accepted: 3 September 2019 / Published: 5 September 2019
(This article belongs to the Special Issue Remote Sensing and Geoscience Information Systems in Applied Sciences)

Abstract

:
Spatiotemporal fusion methods provide an effective way to generate both high temporal and high spatial resolution data for monitoring dynamic changes of land surface. But existing fusion methods face two main challenges of monitoring the abrupt change events and accurately preserving the spatial details of objects. The Flexible Spatiotemporal DAta Fusion method (FSDAF) was proposed, which can monitor the abrupt change events, but its predicted images lacked intra-class variability and spatial details. To overcome the above limitations, this study proposed a comprehensive and automated fusion method, the Enhanced FSDAF (EFSDAF) method and tested it for Landsat–MODIS image fusion. Compared with FSDAF, the EFSDAF has the following strengths: (1) it considers the mixed pixels phenomenon of a Landsat image, and the predicted images by EFSDAF have more intra-class variability and spatial details; (2) it adjusts the differences between Landsat images and MODIS images; and (3) it improves the fusion accuracy in the abrupt change area by introducing a new residual index (RI). Vegetation phenology and flood events were selected to evaluate the performance of EFSDAF. Its performance was compared with the Spatial and Temporal Adaptive Reflectance Fusion Model (STARFM), the Spatial and Temporal Reflectance Unmixing Model (STRUM), and FSDAF. Results show that EFSDAF can monitor the changes of vegetation (gradual change) and flood (abrupt change), and the fusion images by EFSDAF are the best from both visual and quantitative evaluations. More importantly, EFSDAF can accurately generate the spatial details of the object and has strong robustness. Due to the above advantages of EFSDAF, it has great potential to monitor long-term dynamic changes of land surface.

1. Introduction

High spatial resolution remote sensing images with highly frequent observations are significant for monitoring dynamic changes of land surface [1,2,3,4], especially for monitoring short-term and abrupt change events such as floods and forest fires [5,6]. However, due to technical and budget limitations, existing single satellite sensors cannot simultaneously generate high temporal and high spatial resolution images [2,7,8]. For example, the remote sensing images acquired from Landsat, SPOT, and Sentinel 2 satellites have a high spatial resolution (10–30 m), but they have a long coverage period of 5–30 days (hereinafter called “fine resolution” images). The other satellites images (e.g., MODIS and VIIRS) have a daily sampling frequency, but spatial resolution ranges between 250 and 1000 m (hereinafter called “coarse resolution” images). To overcome this constraint, spatiotemporal fusion methods of remote sensing data have been developed to synthesize high spatial and temporal resolution images for monitoring the dynamic changes of land surface by fusing coarse resolution images and fine resolution images [9,10]. In the past decade, the synthetic high spatiotemporal resolution images have been widely used in vegetation phenology monitoring [9,11], urban surface temperatures [12,13,14], urbanization [15], crop yield estimating [16,17], and monitoring sudden and short-term change events (e.g., flood) [18].
According to the literature, the existing spatiotemporal fusion methods can be classified into three categories [19,20,21]: weight function-based, unmixing-based, learning-based. Weight function-based methods estimate the fine resolution images by weighting the similar neighboring information of all input images. The Spatial and Temporal Adaptive Reflectance Fusion Model (STARFM) [7] was the earliest proposed and most widely used spatiotemporal fusion method, which blended the Landsat and MODIS imagery to synthesize Landsat-like images by using a weighted function, but STARFM cannot predict heterogeneous landscape. Then, the Spatial Temporal Adaptive Algorithm for mapping Reflectance Change (STAARCH) [2] and the Enhanced STARFM (ESTARFM) [8] were developed to improve STARFM’s performance in disturbance areas and heterogeneous areas, respectively. However, both STAARCH and ESTARFM required two pairs of MODIS and Landsat images for input data, which are less suitable in cloudy places. Unmixing-based methods predict the fine resolution images by unmixing the coarse resolution images based on the spectral mixing theory. The Multi-sensor Multi-resolution Technique (MMT) [22] was originally proposed based on unmixing, and existing fusion methods based on unmixing were an improvement to MMT. Methods, such as the Spatial Temporal Data Fusion Approach (STDFA) [23] and the Spatial and Temporal Reflectance Unmixing Model (STRUM) [24], were typical unmixing-based methods. The advantages of unmixing-based methods have high computational efficiency, which can generate the time series of images with high spatiotemporal resolution. However, the fusion methods based on unmixing only unmix the coarse resolution images, which did not consider mixed pixels phenomenon of fine resolution images, and its predicted fine resolution images lacked intra-class variability and spatial details [19,25]. In addition, most of both the weight function-based methods and unmixing-based methods supposed the land cover will not change between base date and prediction date, which cannot monitor the abrupt change events.
Learning-based methods have been proposed in recent years, which employ the machine learning algorithms to perform feature learning between coarse resolution images and fine resolution images [26], and simulating the fine resolution images based on the structural similarity of the input images. The SParse-representation-based SpatioTemporal reflectance Fusion Model (SPSTFM) [27] was the first learning-based algorithm, which established a correspondence between the fine resolution and coarse resolution images through the dictionary pair learning to generate the predicted fine resolution images. Following SPSTFM, Song [28] developed another dictionary-based algorithm that only required a pair of coarse resolution and fine resolution images. One of the important improvements of the learning-based methods is that it can predict land cover change events. Although many fusion methods of learning based have recently been proposed, such as Error-Bound-regularized SParse coding (EBSPTM) [29], Extreme Learning Machine (ELM) [30], and Hybrid Color Mapping (HCM) [31], the learning-based fusion methods are relatively novel, which have not been widely used [32], In addition, learning-based methods only consider the statistical relationships between the fine resolution images and coarse resolution images instead of any physical properties of remote sensing signals [28], which predicted fine resolution images cannot exactly remain spatial detail feature and shapes of objects.
Previous literature shows that existing fusion methods have two main improvements including monitoring the abrupt change events and accurately capturing the spatial details of land surface. To solve the problems, hybrid methods were proposed in recent years, and the Flexible Spatiotemporal DAta Fusion method (FSDAF) [32] was a typical hybrid method. FSDAF was based on spectral unmixing analysis and a thin plate spline (TPS) interpolator and combined the weighted function to predict fine resolution images. Because the information of land cover change can be captured by using a TPS interpolator for the MODIS images at the prediction date, the FSDAF can monitor abrupt change events and has been widely used recently. However, the FSDAF did not consider mixed pixels for Landsat images, and its predicted fine resolution images lacked intra-class variability and spatial details. Specifically, the FSDAF has the following three limitations: (1) it directly performs the hard-classification for Landsat image, and its predicted Landsat-like images lack intra-class variability and spatial details; (2) it directly assigns the temporal changes of coarse resolution images to fine resolution images without considering the differences between Landsat and MODIS images; and (3) it introduces a homogeneity index (HI) to guide the residual distribution, which is derived from the classification map at the base date. The HI will not be suitable for guiding the residual distribution when there are land cover changes and misclassifications [33]. In addition, more input parameters were required to be set before the operation of FSDAF, which increased the computational complexity.
To solve the above limitations of the FSDAF method, this study developed a comprehensive and automated fusion method, an enhanced FSDAF (EFSDAF), and tested it by fusing the Landsat images and MODIS images. Compared with FSDAF, EFSDAF has the following strengths: (1) it considers the mixed pixels in Landsat images and uses a globally representative spectral linear mixture model (SVD) of the Landsat pixels for spectral unmixing; (2) it adjusts the differences between Landsat images and MODIS images using a linear model; and (3) it proposes a new residual index (RI) to guide the residual distribution based on the interpolator results of MODIS images at the base date and prediction date. In addition, EFSDAF is an automated spatiotemporal fusion method that does not require additional input parameters compared to FSDAF. In this study, we tested the EFSDAF in the areas of vegetation phenology and flood events and compared it with three other popular fusion methods: STARFM, STRUM, and FSDAF.

2. Method

2.1. Definitions and Notations

Before introducing the details of EFSDAF, some important definitions and notations are given here for convenience and clarity.
T B and T P define the base date and the prediction date, respectively. S 2 is the number of fine resolution pixels within one coarse resolution pixel. m represents the m-th endmembers within one coarse resolution pixel or one fine resolution pixel. x i , y i and x i j , y i j denote the coordinate index of the i-th coarse resolution pixels and the j-th fine pixels within the i-th coarse pixels, respectively, and j = 1 … s 2 . The coarse resolution images observed at T B and T P are stored in C B x i ,   y i , b and C P x i , y i , b , and the fine resolution images observed at T B and T P are stored in F B x i j , y i j , b and F P x i j , y i j , b . b is the number of bands for coarse resolution images and fine resolution images. A F B x i j , y i j , m and A F P x i j , y i j , m denote the abundances of the fine resolution pixels at T B and T P , and A C B x i , y i , m and A C P x i , y i , m denote the abundances of the coarse resolution pixels at T B and T P , respectively.

2.2. Theoretical Basis of EFSDAF

Like FSDAF, the input images of EFSDAF include, a MODIS image at T B , a Landsat image at T B and a MODIS image at T P . The output of EFSDAF is a predicted fine resolution image (e.g., Landsat-like) at T P . Unlike FSDAF, EFSDAF considers more than one land cover type in both the Landsat images (30 m) and MODIS images (500 m). Therefore, the endmembers determination and abundances extraction of the Landsat image at T B are firstly carried out in EFSDAF. EFSDAF includes five main steps: (1) endmember determination and spectral unmixing for Landsat image at T B ; (2) temporal prediction ( F T P ) for no land cover change from T B to T P ; (3) spatial prediction ( F S P ) for land cover change at T P ; (4) residual distribution by using a new residual index (RI); and (5) final prediction of a Landsat-like image using neighborhood in a sliding window. The workflow of EFSDAF is outlined by the flowchart in Figure 1. Detailed descriptions of each step in EFSDAF are given below. The code of EFSDAF can be found at the URL https://github.com/max19951001/EFSDAF.git.

2.2.1. Endmember Determination and Spectral Unmixing for Landsat Images at T B

Endmember extraction and spectral unmixing of Landsat images at T B are the first step and key step for EFSDAF in this paper due to the heterogeneity of land surface with more than one land cover type in Landsat pixels. In this study, a globally representative spectral linear mixture model (SVD model) [34,35,36] shared by Small for Landsat images was selected as endmembers that were classified into three types: substrate (S), vegetation (V), and dark surface (D) [20], and as a linear method, the fully constrained least squares (FCLS) [37] method was applied to the spectral unmixing. The abundance A F B x i j , y i j , m calculated by FCLS varies from 0 to 1, and the sum of A F B x i j , y i j , m is 1.

2.2.2. Temporal Prediction ( F T P ) for No Land Cover Change from T B to T P

  • Temporal changes of each endmember at the MODIS pixels
The temporal prediction of EFSDAF assumes no land cover change from T B to T P . In other words, the endmembers and abundances of Landsat pixels will not change from T B to T P . According to the linear mixture model [38], the values of Landsat pixels are a linear mixture of endmembers’ values and abundances. Hence, the values of Landsat pixels at T B and T P can be expressed as:
F B x i j , y i j , b = m = 1 n m A F B x i j , y i j , m × R F B m , b + φ
F p x i j , y i j , b = m = 1 n m A F p x i j , y i j , m × R F p m , b + φ
where n m defines the number of endmembers and m is the m-th endmembers for one Landsat pixel, R F B m , b and R F P m , b denote the reflectance values of each endmember for all Landsat image bands at T B and T P , respectively. φ is the system error. According to the above assumption, the endmembers and abundances of Landsat pixels will not change from T B to T P , i.e., A F B = A F P , and φ is constant. From Equations (1) and (2), we have:
Δ F x i j , y i j , b = m = 1 n m A F B x i j , y i j , m × Δ R F m , b
where Δ F denotes the temporal changes of the Landsat images. Δ R F are the reflectance changes of each endmember for all Landsat image bands. Among the variables above, if Δ R F is known, Δ F can be calculated by a linear mixture model. Therefore, the key is to solve Δ R F .
Similar to the temporal changes of Landsat images, the temporal changes of MODIS images from T B to T P can be expressed as:
Δ C x i , y i , b = m = 1 n m A C B x i , y i , m × Δ R C m , b .
where Δ C denotes the temporal changes of the MODIS images. Δ R C are the reflectance changes of each endmember for all MODIS image bands. Among the variables above, Δ C could be calculated through Δ C = C P C B . Due to the previous assumption of no land cover change from T B to T P , the types and spatial distributions of the endmembers should be the same for Landsat and MODIS images in the same regions. Hence, A C B can be aggregated from A F B with:
A C B x i , y i , m = 1 s 2 k = 1 s 2 A F B x i j , y i j , m
where A C B x i , y i , m is calculated by averaging the endmember abundances of all Landsat pixels in one MODIS pixel. However, A C B x i , y i , m is the ratio of the number of Landsat pixels for class m to s 2 in FSDAF. Although Δ C x i , y i , b and A C B x i , y i , m from Equations (4) and (5) are known variables, Equation (4) cannot be solved because there are n m unknown values of Δ R C m , b .
In this study, assuming that the land cover types are the same for a small area according to the first law of geography [39] stating that near things are more related than distant things, it is reasonable that Δ R C m , b are the same in a small area [20]. However, FSDAF assumes that the temporal changes of each class are the same among all MODIS pixels, which does not match the actual situation. A sliding window of MODIS pixels is introduced to perform spatial unmixing in this study, and the size of the window should be larger than the number of endmembers [40]. The equations [32] are as follows:
Δ C x 1 , y 1 , b Δ C x i , y i , b Δ C x n , y n , b = A C B x 1 , y 1 , 1   A C B x 1 , y 1 , 2     A C B x 1 , y 1 , m         A C B x i , y i , 1   A C B x i , y i , 2     A C B x i , y i , m         A C B x n , y n , 1   A C B x n , y n , 2     A C B x n , y n , m Δ R C 1 , b Δ R C l , b Δ R C m , b
where Δ R C m , b can be calculated by computing a least squares best fit solution. Selecting k of the purest MODIS pixels of each endmember in the sliding window for least squares best fit aims to reduce the errors by collinearity [32]. The final determined MODIS pixels for least squares best fit should be smaller than the size of the sliding window and larger than the number of endmembers.
  • Adjustment of the differences between the Landsat images and MODIS images
Theoretically speaking, the reflectance values of the Landsat pixels corresponding to the MODIS pixels are the same. However, due to the physical differences of the sensors, the effects of bandwidth, weather conditions, and atmospheric correction, differences between MODIS images and Landsat images are inevitable. As with previous studies [20,41], a linear model is introduced in this study to adjust the differences between Landsat images and MODIS images.
R F B m , b = a 1 b × R C B m , b + b 1
R F P m , b = a 1 b × R C P m , b + b 1
where a 1 b and b 1 are the slope and interception of the linear model, respectively. From Equations (7) and (8), the differences between Δ R C m , b and Δ R F m , b can be calculated as the following Equation (9).
Δ R F m , b = a 1 b × Δ R C m , b
Δ R F m , b can be obtained from Equations (5)–(9) in this study. However, the Δ R F m , b of FSDAF were calculated by assigning the temporal changes of MODIS images to Landsat images without considering the differences of Landsat and MODIS images.
  • Temporal prediction for the fine resolution image at T P
Δ F is linearly mixed with A F B from Equation (3), and the final temporal prediction can be calculated by Equation (10) as follows:
F T P x i j , y i j , b = F B x i j , y i j , b + Δ F x i j , y i j , b
Although the final temporal prediction can be calculated by Equation (10), all of the above calculations are based on the assumption of no land cover change from T B to T P . Hence, the temporal prediction cannot accurately predict abrupt change events from T B to T P . In addition, because this prediction only uses the information of the temporal changes between T B and T P , we call Section 2.2.2 the temporal prediction.

2.2.3. Spatial Prediction ( F S P ) for Land Cover Change at T P

  • Analysis and calculation of the residual
The temporal prediction from Section 2.2.2 is not an accurate prediction when there are land cover changes. Specifically, there is a certain deviation between the true values ( F p ) and the temporal prediction values ( F T P ). This study introduces a residual R between the F P and the F T P as follows:
1 S 2 s = 1 s 2 F P x i j , y i j , b = 1 S 2 s = 1 s 2 F T P x i j , y i j , b + R x i , y i , b
Furthermore, the value of one MODIS pixel is equal to the average of all of Landsat pixels within one MODIS pixels and a system deviation φ [7]. Hence, the values of MODIS pixels at T B and T P can be expressed as:
C B x i , y i , b = 1   s 2 s = 1 s 2 F B x i j , y i j , b + φ
C P x i , y i , b = 1 s 2 s = 1 s 2 F P x i j , y i j , b + φ
From Equations (11)–(13), we have:
R x i , y i , b = Δ C x i , y i , b 1 S 2 s = 1 s 2 F T P x i j , y i j , b s = 1 s 2 F B x i j , y i j , b
As can be seen from the above Equations (11)–(14), the calculation and distribution of the residual R is a key step for obtaining the final prediction at T P .
  • Spatial prediction of the MODIS image at T P
As described above, the source of the residual R is in the area of land cover change. However, the information of land cover change can only be obtained from the MODIS image at T P . Therefore, downscaling the MODIS image to Landsat image level is the vital step to obtain the information of the land cover change at T P . In EFSDAF, a Thin Plate Spline (TPS) interpolation method is applied to downscale the MODIS image at T P . TPS is a spatial interpolation technique for point data based on spatial dependence [42], which was used to the spatiotemporal fusion method because it has high interpolation accuracy [32,43]. The interpolation result of the MODIS image at T P will be marked as F P S P x i j , y i j , b , which is another prediction of the Landsat image at T P . Unlike FSDAF, the MODIS image at T B is also interpolated using the TPS method to calculate the residual index (RI) in Section 2.2.4 for EFSDAF, and its result will be marked as F B S P x i j , y i j , b . Because the spatial information is obtained from MODIS images using the TPS interpolator method, we call Section 2.2.3 spatial prediction.

2.2.4. Residual Distribution by Using a New Residual Index (RI)

From Equation (11), the predicted Landsat-like image is calculated by adding a reasonable residual based on the temporal prediction. Therefore, the residual distribution is the key step for the final prediction. In FSDAF, a homogeneity index (HI) derived from the classification map of the Landsat image at T B was introduced to guide the residual distribution. Because the HI only depends on the classification map at T B , it will be unreasonable when the land cover type has changed from T B to T P or the classification map is wrong [33]. Hence, a new residual index (RI) is proposed to reasonably guide the residual distribution in this study, and more residual will be assigned to the area of land cover change. According to Section 2.2.3, the spatial variation Δ F B P x i j , y i j , b can be expressed as:
Δ F B P x i j , y i j , b = F P S P x i j , y i j , b F B S P x i j , y i j , b
where Δ F B P x i j , y i j , b defines the changes of interpolation results of MODIS images from T B to T P , and the larger Δ F B P x i j , y i j , b , the larger the changes of land cover type. In EFSDAF, we assume the changes of land cover type occur in a small area. If Δ F B P x i j , y i j , b in one Landsat pixel is greater than the average of the Δ F B P , and we think the Landsat pixel x i j , y i j is the area of land cover change. The RI can be calculated as follows:
R I x i j , y i j , b = k = 1 n I k n w
where I k is equal to 1 when the k-th Landsat pixels is determined to the pixel of land cover change within a sliding window; otherwise, I k is equal to 0. n w defines the window size. The RI is the ratio of the numbers of I k = 1 in the window to n w . The range of RI is 0–1, and a larger RI indicates more residual to the area of land cover change.
Then we use the same weighting function to distribute the final residual as FSDAF. The weighting function is defined as follow:
C W x i j , y i j , b = E h e x i j , y i j , b × R I x i j , y i j , b + E h o x i j , y i j , b × 1 R I x i j , y i j , b
where E h e x i j , y i j , b = R x i , y i , b , and E h e x i j , y i j , b is the residual for the area of land cover change. E h o x i j , y i j , b = F P S P x i j , y i j , b F T P x i j , y i j , b , and E h o x i j , y i j , b is the residual for area of no land cover change.
The final residual r x i j , y i j , b can be calculated as follows:
r x i j , y i j , b = s 2 × R x i , y i , b × W x i j , y i j , b
where W x i j , y i j , b is the normalized C W x i j , y i j , b .
According to Equations (10) and (11), the final changes of the Landsat images between T B and T P can be obtained from Equation (19):
Δ F f x i j , y i j , b = r x i j , y i j , b + Δ F x i j , y i j , b
where Δ F f x i j , y i j , b and Δ F x i j , y i j , b are the final changes of Landsat images and the changes of temporal prediction from T B to T P , respectively.

2.2.5. Final Prediction of the Landsat-Like Image Using Neighborhood in a Sliding Window

Theoretically, the final prediction of Landsat-like image F t p x i j , y i j , b at T P can be obtained by the following equation.
F t p x i j , y i j , b = F B x i j , y i j , b + Δ F f x i j , y i j , b
However, uncertain errors are inevitable in EFSDAF because of the complexity of the calculation process and the characteristics of pixel-by-pixel. In addition, the final prediction will lead to block effects because the residual distribution is for the MODIS pixels. In this study, EFSDAF adopts the same solution as STARFM, which uses the information of neighborhood to reduce the uncertainties and smooth final prediction in a sliding window. The final predicted Landsat-like image F f p x i j , y i j , b can be calculated by Equation (21), and detailed information can be found in STARFM [7].
F f p x i j , y i j , b = F B x i j , y i j , b + k = 1 n w k × Δ F f x i j , y i j , b
where n is the number of similar pixels for Landsat pixels in a sliding window, and w k is the weight for each similar pixel.

3. Testing Experiment

3.1. Study Area and Data

The EFSDAF is proposed to accurately monitor dynamic changes of land surface including gradual change (e.g., vegetation phenology) and abrupt change events (e.g., flood). In this study, vegetation phenology and flood events are selected to evaluate the performance of EFSDAF.
The selected two datasets were from the datasets shared by Emelyanova [10], which were collected in the Lower Gwydir Catchment in Northern New South Wales, Australia (149.2815° E, 29.0855° S, hereinafter called “Gwydir”) and have been widely applied in the evaluation of spatiotemporal fusion methods [10,20,21,32,44]. All images were atmospherically corrected and geographically co-registered, which was resampled to 25 m with the nearest neighbor method by Emelyanova. In EFSDAF, we aggregated the MODIS pixels to 500 m resolution in the process of residual distribution. Two pairs of Landsat and MODIS images of the first dataset (vegetation phenology) were acquired on 5 July 2004 and 6 August 2004 in Gwydir, and the size of the selected images is 1400 × 1400   pixels . The first dataset was a typical heterogeneous region, and temporal dynamic changes were mainly caused by phenological change from 5 July to 6 August 2004. The selected Landsat and MODIS images of the first dataset are shown in Figure 2. The second dataset (flood) was acquired on 11 November 2004 and 12 December 2004, and the size of selected images is 800 × 800 pixels. A large flood occurred from 26 November 2004 to 12 December 2004, and the dataset was a typical abrupt change area. The selected Landsat and MODIS images of the second dataset are shown in Figure 3.

3.2. Comparison and Evaluation of EFSDAF with STARFM, STRUM, and FSDAF

In this study, we compared EFSDAF with three other fusion methods: STARFM, STRUM, and FSDAF. Because these three fusion methods have been widely used, and each of them only needs one pair of Landsat and MODIS images at T B and one MODIS image at T P as input data.
All fusion images are qualitatively and quantitatively evaluated by comparing the predicted Landsat-like image with true images at T P . The qualitative evaluation is to visually compare the predicted images with the true images, and the differences between the predicted images of the four fusion methods are magnified and highlighted by using yellow elliptical circles and black squares. Representative assessment indexes as quantitative evaluation are used to objectively evaluate the predicted images and true images pixel by pixel, and they are Average Difference (AD), Root Mean Square Error (RMSE), Correlation Coefficient (CC), and Structure Similarity (SSIM) [45], respectively. The evaluation criteria are as follows: the closer AD and RMSE are to 0, the closer the predicted value to the true value; the closer CC and SSIM are to 1, the more similar the predicted images are to the true images.

4. Results

4.1. Experiment in Gradual Change Area (Vegetation Phenology)

From Figure 4, the predicted Landsat-like images by four methods can monitor the changes of vegetation growth, and the results of the four fusion methods are similar to the true image on the whole. However, the yellow elliptical circles in Figure 4 show the differences of the four methods, and we can get that STARFM blurs the boundaries of different objects, STRUM produces block effects, FSDAF cannot retain the spatial details, and EFSDAF is closest to the true image. Especially, EFSDAF can accurately retain the boundaries of objects from Figure 5, and it can capture more spatial details of the object than FSDAF.
From quantitative indices of four methods in Table 1, the fusion results by four methods can achieve high accuracy. For all six bands and the mean, it can be seen that the accuracy of EFSDAF is the highest by comparing it with the other three methods, which have lower AD and RMSE, as well as higher CC and SSIM. Taking the mean of all bands as an example, the CC values of the four methods are 0.807, 0.826, 0.858, and 0.883 from left to right, respectively. In addition, because the vegetation phenology change happened in dataset 1, the prediction accuracy of band 4 (NIR) can better illustrate the performance of the four methods. From the scatter plot of Figure 6, it can be confirmed that the synthetic image values by EFSDAF are closer to the true values than the other three methods.

4.2. Experiment in Abrupt Change Area (Flood)

From the visual comparison, only FSDAF and EFSDAF can capture the information of land cover change in Figure 7, it can be more clearly and accurately seen from the zoom-in black area in Figure 8. In addition, the spatial details restored by EFSDAF are most complete and most similar with the true image from the yellow elliptical circles of Figure 7. From the yellow elliptical circles of Figure 8, the small flood change information can be accurately captured by EFSDAF, but it cannot be captured by FSDAF.
Comparing the results of the quantitative indices in Table 2, STARFM and STRUM present similar results, and EFSDAF is better than FSDAF (CC 0.808 vs. 0.827, SSIM 0.796 vs. 0.815 for band 4). Furthermore, since the water body information shows lower reflectivity for band 4 (NIR), we compare the scatter plots of band 4 for the four methods in Figure 9. It can be concluded that FSDAF and EFSDAF are far superior to STARFM and STRUM, and EFSDAF is better than FSDAF.

5. Discussion

The proposed EFSDAF method in this paper was an improvement on FSDAF. EFSDAF considered the mixed pixels of both Landsat images and MODIS images, and it predicted the Landsat-like images by combining spectral unmixing analysis and TPS interpolator. The experiment results showed that the fusion results by EFSDAF were better than FSDAF.

5.1. Improvements of EFSDAF Compared with FSDAF

The fusion results of EFSDAF were better than FSDAF due to the following reasons. Firstly, EFSDAF considered the mixed pixels phenomenon of the fine resolution image (e.g., Landsat), which was the most significant improvement of EFSDAF. In this study, the SVD model was introduced to perform the spectral unmixing by using a fully constrained least squares method. Moreover, the abundances of MODIS pixels were calculated by averaging the abundances of all Landsat pixels in one MODIS pixel, and the final changes of Landsat images were solved by a linear mixture that combined the abundances of endmembers and the variations of endmembers for Landsat images from T B to T p . Hence, the proposed EFSDAF considered the intra-class variability, and it can reserve more spatial details than FSDAF, which can be seen in the yellow elliptical circles of Figure 4 and Figure 7. Secondly, the spatial unmixing of the coarse resolution image (e.g., MODIS) was performed in a sliding window, which was different from the spatial unmixing of FSDAF using the whole MODIS image. It was more in line with the actual situation by using sliding windows for spatial unmixing due to the heterogeneity of land surface. The size of the sliding window for spatial unmixing was recommended to 7 × 7 MODIS pixels because the fusion accuracy is the best. Furthermore, selecting k purest MODIS pixels of a sliding window for spatial unmixing ensured the minimal impacts of collinearity [32]. Thirdly, the differences between the Landsat and MODIS images were considered, which included the differences of the sensors and the pre-processing errors. In this study, we used a linear model to adjust the differences between the Landsat and MODIS images, but FSDAF did not consider the differences. Finally, a new residual index (RI) was introduced to guide the residual distribution instead of the homogeneity index (HI) of FSDAF. The HI will not be suitable to guide the residual distribution when there are land cover changes or misclassification [33] because it was calculated from the classification map at T B . The proposed RI adequately considered the source of the errors, and it was based on the differences of the interpolator results between the MODIS image at T B and the MODIS image at T P . Greater differences represent more significant land cover change and will have a larger RI value. The RI was calculated in a sliding window, and the size of the window should not be too large because the result will not be accurate for the edge of different land cover types. The recommended size was 51 × 51 Landsat pixels according to Section 2.2.5 in this study. More information of land cover change can be monitored by EFSDAF from the yellow elliptical circles in Figure 8e,f. In addition, EFSDAF is an automated fusion method compared to FSDAF because additional input parameters were not required.

5.2. Influence of Endmember Variability on EFSDAF

The endmember variability directly affected the abundances of the Landsat image which may further affect the final fusion precision of EFSDAF. Hence, we evaluated the performance of EFSDAF by using different endmember numbers. In this study, the SVD model shared by Small and 3~5 endmembers extracted from the Landsat image at T B were selected to test the influence of endmember variability on EFSDAF, and the final results of the two datasets were shown in Table 3 below (three bands for testing, NIR-red-green). As can be seen from Table 3, the predicted results of EFSDAF using four different endmembers are close in the two datasets. Therefore, the influence of endmember variability on the final fusion results can be negligible, and EFSDAF is robust to endmember variability. For the automated operation of EFSDAF, the SVD model of Landsat image was used to perform the spectral unmixing in this study.

5.3. The Effect of Input Images on the Predicted Values of EFSDAF

Like FSDAF, the minimum input images were used to predict the final Landsat-like images in EFSDAF. Hence, the fusion accuracy of EFSDAF was directly influenced by the input images [10,19]. Particularly, the spatial information and temporal information of the input images would be fully utilized in EFSDAF. Figure 10 showed the differences between the predicted Landsat-like image of EFSDAF and input images in dataset 2. From Figure 10b–d, it can be seen that the land cover type has changed, and the information of land cover change can be captured by comparing Figure 10b with Figure 10c. However, there was an obvious difference between the predicted image in the yellow ellipse area of Figure 10b and the true image (Figure 10a). After comparing and analyzing the input images and the predicted image in Figure 10a,d, it can be concluded that the input MODIS image at T P lacks the information of land cover change in the yellow ellipse from Figure 10c; therefore, the final predicted image in Figure 10a was different from the true image at T P from Figure 10a. However, comparing the abrupt change area of MODIS images from T B to T P in Figure 10c,d, EFSDAF can completely capture the spatial details in Figure 10b. In other words, EFSDAF can only accurately monitor the information of abrupt change in MODIS images. Hence, the quality of the input images is crucial to the fusion accuracy of EFSDAF.

5.4. Applications of EFSDAF to other Remote Sensing Products and Sensors

The EFSDAF in this study was tested using surface reflectance values of MODIS and Landsat images, but the purpose of EFSDAF is to study various complex surface dynamics. Therefore, EFSDAF can be applied to vegetation monitoring (e.g., NDVI), crop growth, and other sudden events such as urbanization, forest fires, and landslides. In addition, with the frequent launch of satellites for the Earth Observation System in recent years, existing satellite sensors such as Sentinel 2 MSI (10, 20, 60 m) images can replace Landsat images as fine resolution images, and like Sentinel 3 OLCI (300 m), NPP-VIIRS images (375/750 m) can replace MODIS images as coarse resolution images. We are looking forward to testing EFSDAF for other applications as well as satellite sensors.

5.5. Limitations of EFSDAF

EFSDAF can predict both gradual change and abrupt change events, but it cannot accurately monitor tiny changes (e.g., small flood information in this study) because tiny changes are not shown in one MODIS pixel. In other words, the changes of 20 × 20 Landsat pixels can be displayed in one MODIS pixel in this study. Therefore, the prediction accuracy of tiny changes can be improved by blending Landsat images (30 m) and Sentinel 3 OLCI images (300 m), but the predicted images still cannot monitor all actual changes. In addition, the computational efficiency of EFSDAF is the slowest compared with other three methods due to the complexity of the calculation process. However, the computational efficiency of fusion methods will be ignored in the future with the rapid development of high-performance computing platforms like Google Earth Engine (GEE) [46].

6. Conclusions

In order to improve the shortcomings of FSDAF, this study developed the Enhanced FSDAF (EFSDAF) to accurately monitor dynamic changes of land surface. EFSDAF improved FSDAF in the following aspects: (1) unlike existing spatiotemporal fusion methods of unmixing based (e.g., STRUM, FSDAF), EFSDAF considered the mixed pixels of fine resolution images (Landsat), and the fusion results by EFSDAF takes into account intra-class variability and can accurately reserve the spatial details of land surface; (2) the differences between the coarse resolution images and the fine resolution images were adjusted; (3) a new residual index (RI) was proposed to guide the residual distribution, which improved the fusion accuracy in the abrupt change area. Experimental results demonstrated that:
(1)
EFSDAF can accurately monitor both the gradual change and abrupt change events. More importantly, EFSDAF can reserve more spatial details of land surface and has a stronger robustness than FSDAF.
(2)
EFSDAF can monitor more information of land cover change than FSDAF by introducing a new residual index (RI) to guide residual distribution because the proposed RI considers the actual source of residual.
(3)
EFSDAF is an automated fusion method because it does not need additional input parameters, which has great potential to monitor long-term dynamic changes of land surface using high spatiotemporal images.
In addition, we also expect that EFSDAF can be applied to other products of remote sensing and other satellite sensors.

Author Contributions

C.S. designed and developed the EFSDAF code and wrote the manuscript. X.W. directed and revised the manuscript. M.Z. and X.L. processed the experimental data. L.N. and H.H. made the Figures and Charts. X.Z. revised the manuscript.

Funding

This study was financially supported by National Natural Science Foundation of China [Grant number 41971387 and Grant number 41071271], Shaanxi Province Natural Science Foundation [Grant number 2015JM4132], Strategic Priority Research Program of Chinese Academy of Sciences, China [Grant number XDA2004030201], and Shaanxi Key Laboratory of Ecology and Environment of River Wetland [Grant number SXSD201701].

Acknowledgments

The authors are grateful to I. Emelyanova for sharing test data in the remote sensing community and thank C. Small for sharing the spectra of Global SVD Endmembers. Thanks to the FSDAF code shared by Xiaolin Zhu and the STRUM code provided by Jianhang Ma. Most of all, I want to thank my girlfriend Qiaona Sun for the constant support and encouragement.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Yang, X.; Lo, C. Using a time series of satellite imagery to detect land use and land cover changes in the Atlanta, Georgia metropolitan area. Int. J. Remote Sens. 2002, 23, 1775–1798. [Google Scholar] [CrossRef]
  2. Hilker, T.; Wulder, M.A.; Coops, N.C.; Linke, J.; McDermid, G.; Masek, J.G.; Gao, F.; White, J.C. A new data fusion model for high spatial-and temporal-resolution mapping of forest disturbance based on Landsat and MODIS. Remote Sens. Environ. 2009, 113, 1613–1627. [Google Scholar] [CrossRef]
  3. Roy, D.P.; Wulder, M.A.; Loveland, T.R.; Woodcock, C.; Allen, R.G.; Anderson, M.C.; Helder, D.; Irons, J.R.; Johnson, D.M.; Kennedy, R. Landsat-8: Science and product vision for terrestrial global change research. Remote Sens. Environ. 2014, 145, 154–172. [Google Scholar] [CrossRef] [Green Version]
  4. Townshend, J.R.; Masek, J.G.; Huang, C.; Vermote, E.F.; Gao, F.; Channan, S.; Sexton, J.O.; Feng, M.; Narasimhan, R.; Kim, D. Global characterization and monitoring of forest cover using Landsat data: opportunities and challenges. Int. J. Digital Earth 2012, 5, 373–397. [Google Scholar] [CrossRef] [Green Version]
  5. Zhang, F.; Zhu, X.; Liu, D. Blending MODIS and Landsat images for urban flood mapping. Int. J. Remote Sens. 2014, 35, 3237–3253. [Google Scholar] [CrossRef]
  6. Filipponi, F. Exploitation of Sentinel-2 Time Series to Map Burned Areas at the National Level: A Case Study on the 2017 Italy Wildfires. Remote Sens. 2019, 11, 622. [Google Scholar] [CrossRef]
  7. Gao, F.; Masek, J.; Schwaller, M.; Hall, F. On the blending of the Landsat and MODIS surface reflectance: Predicting daily Landsat surface reflectance. IEEE Trans. Geosci. Remote Sens. 2006, 44, 2207–2218. [Google Scholar]
  8. Zhu, X.; Chen, J.; Gao, F.; Chen, X.; Masek, J.G. An enhanced spatial and temporal adaptive reflectance fusion model for complex heterogeneous regions. Remote Sens. Environ. 2010, 114, 2610–2623. [Google Scholar] [CrossRef]
  9. Hilker, T.; Wulder, M.A.; Coops, N.C.; Seitz, N.; White, J.C.; Gao, F.; Masek, J.G.; Stenhouse, G. Generation of dense time series synthetic Landsat data through data blending with MODIS using a spatial and temporal adaptive reflectance fusion model. Remote Sens. Environ. 2009, 113, 1988–1999. [Google Scholar] [CrossRef]
  10. Emelyanova, I.V.; McVicar, T.R.; Van Niel, T.G.; Li, L.T.; van Dijk, A.I. Assessing the accuracy of blending Landsat–MODIS surface reflectances in two landscapes with contrasting spatial and temporal dynamics: A framework for algorithm selection. Remote Sens. Environ. 2013, 133, 193–209. [Google Scholar] [CrossRef]
  11. Gao, F.; Hilker, T.; Zhu, X.; Anderson, M.; Masek, J.; Wang, P.; Yang, Y. Fusing Landsat and MODIS data for vegetation monitoring. IEEE Geosci. Remote Sens. Mag. 2015, 3, 47–60. [Google Scholar] [CrossRef]
  12. Zhao, Z.-Q.; He, B.-J.; Li, L.-G.; Wang, H.-B.; Darko, A. Profile and concentric zonal analysis of relationships between land use/land cover and land surface temperature: Case study of Shenyang, China. Energy Build. 2017, 155, 282–295. [Google Scholar] [CrossRef]
  13. Wu, P.; Shen, H.; Zhang, L.; Göttsche, F.-M. Integrated fusion of multi-scale polar-orbiting and geostationary satellite observations for the mapping of high spatial and temporal resolution land surface temperature. Remote Sens. Environ. 2015, 156, 169–181. [Google Scholar] [CrossRef]
  14. Weng, Q.; Fu, P.; Gao, F. Generating daily land surface temperature at Landsat resolution by fusing Landsat and MODIS data. Remote Sens. Environ. 2014, 145, 55–67. [Google Scholar] [CrossRef]
  15. Wang, J.; Huang, B. A spatiotemporal satellite image fusion model with autoregressive error correction (AREC). Int. J. Remote Sens. 2018, 39, 6731–6756. [Google Scholar] [CrossRef]
  16. Gao, F.; Anderson, M.; Daughtry, C.; Johnson, D. Assessing the variability of corn and soybean yields in central Iowa using high spatiotemporal resolution multi-satellite imagery. Remote Sens. 2018, 10, 1489. [Google Scholar] [CrossRef]
  17. Wang, Q.; Atkinson, P.M. Spatio-temporal fusion for daily Sentinel-2 images. Remote Sens. Environ. 2018, 204, 31–42. [Google Scholar] [CrossRef] [Green Version]
  18. Huang, C.; Chen, Y.; Zhang, S.; Liu, R.; Shi, K.; Li, L.; Wu, J. Blending NPP-VIIRS and Landsat OLI Images for Flood Inundation Monitoring; CSIRO: Canberra, Australia, 2015. [Google Scholar]
  19. Zhu, X.; Cai, F.; Tian, J.; Williams, T. Spatiotemporal fusion of multisource remote sensing data: literature survey, taxonomy, principles, applications, and future directions. Remote Sens. 2018, 10, 527. [Google Scholar]
  20. Ma, J.; Zhang, W.; Marinoni, A.; Gao, L.; Zhang, B. An improved spatial and temporal reflectance unmixing model to synthesize time series of landsat-like images. Remote Sens. 2018, 10, 1388. [Google Scholar] [CrossRef]
  21. Chen, B.; Huang, B.; Xu, B. Comparison of spatiotemporal fusion models: A review. Remote Sens. 2015, 7, 1798–1835. [Google Scholar] [CrossRef]
  22. Zhukov, B.; Oertel, D.; Lanzl, F.; Reinhackel, G. Unmixing-based multisensor multiresolution image fusion. IEEE Trans. Geosci. Remote Sens. 1999, 37, 1212–1226. [Google Scholar] [CrossRef]
  23. Wu, M.; Niu, Z.; Wang, C.; Wu, C.; Wang, L. Use of MODIS and Landsat time series data to generate high-resolution temporal synthetic Landsat data using a spatial and temporal reflectance fusion model. J. Appl. Remote Sens. 2012, 6, 063507. [Google Scholar]
  24. Gevaert, C.M.; García-Haro, F.J. A comparison of STARFM and an unmixing-based algorithm for Landsat and MODIS data fusion. Remote Sens. Environ. 2015, 156, 34–44. [Google Scholar] [CrossRef]
  25. Belgiu, M.; Stein, A. Spatiotemporal image fusion in remote sensing. Remote Sens. 2019, 11, 818. [Google Scholar] [CrossRef]
  26. Chen, B.; Huang, B.; Xu, B. A hierarchical spatiotemporal adaptive fusion model using one image pair. Int. J. Digital Earth 2017, 10, 639–655. [Google Scholar] [CrossRef]
  27. Huang, B.; Song, H. Spatiotemporal reflectance fusion via sparse representation. IEEE Trans. Geosci. Remote Sens. 2012, 50, 3707–3716. [Google Scholar] [CrossRef]
  28. Song, H.; Huang, B. Spatiotemporal satellite image fusion through one-pair image learning. IEEE Trans. Geosci. Remote Sens. 2012, 51, 1883–1896. [Google Scholar] [CrossRef]
  29. Wu, B.; Huang, B.; Zhang, L. An error-bound-regularized sparse coding for spatiotemporal reflectance fusion. IEEE Trans. Geosci. Remote Sens. 2015, 53, 6791–6803. [Google Scholar] [CrossRef]
  30. Liu, X.; Deng, C.; Wang, S.; Huang, G.-B.; Zhao, B.; Lauren, P. Fast and accurate spatiotemporal fusion based upon extreme learning machine. IEEE Geosci. Remote Sens. Lett. 2016, 13, 2039–2043. [Google Scholar] [CrossRef]
  31. Kwan, C.; Budavari, B.; Gao, F.; Zhu, X. A hybrid color mapping approach to fusing MODIS and landsat images for forward prediction. Remote Sens. 2018, 10, 520. [Google Scholar] [CrossRef]
  32. Zhu, X.; Helmer, E.H.; Gao, F.; Liu, D.; Chen, J.; Lefsky, M.A. A flexible spatiotemporal method for fusing satellite images with different resolutions. Remote Sens. Environ. 2016, 172, 165–177. [Google Scholar] [CrossRef]
  33. Liu, M.; Yang, W.; Zhu, X.; Chen, J.; Chen, X.; Yang, L.; Helmer, E.H. An Improved Flexible Spatiotemporal DAta Fusion (IFSDAF) method for producing high spatiotemporal resolution normalized difference vegetation index time series. Remote Sens. Environ. 2019, 227, 74–89. [Google Scholar] [CrossRef]
  34. Sousa, D.; Small, C. Global cross-calibration of Landsat spectral mixture models. Remote Sens. Environ. 2017, 192, 139–149. [Google Scholar] [CrossRef] [Green Version]
  35. Small, C.; Milesi, C. Multi-scale standardized spectral mixture models. Remote Sens. Environ. 2013, 136, 442–454. [Google Scholar] [CrossRef] [Green Version]
  36. Small, C. The Landsat ETM+ spectral mixing space. Remote Sens. Environ. 2004, 93, 1–17. [Google Scholar] [CrossRef]
  37. Heinz, D.C. Fully constrained least squares linear spectral mixture analysis method for material quantification in hyperspectral imagery. IEEE Trans. Geosci. Remote Sens. 2001, 39, 529–545. [Google Scholar] [CrossRef] [Green Version]
  38. Ma, W.-K.; Bioucas-Dias, J.M.; Chan, T.-H.; Gillis, N.; Gader, P.; Plaza, A.J.; Ambikapathi, A.; Chi, C.-Y. A signal processing perspective on hyperspectral unmixing: Insights from remote sensing. IEEE Signal Proces. Mag. 2013, 31, 67–81. [Google Scholar] [CrossRef]
  39. Steven, M.D.; Malthus, T.J.; Baret, F.; Xu, H.; Chopping, M.J. Intercalibration of vegetation indices from different sensor systems. Remote Sens. Environ. 2003, 88, 412–422. [Google Scholar] [CrossRef]
  40. Zurita-Milla, R.; Kaiser, G.; Clevers, J.; Schneider, W.; Schaepman, M. Downscaling time series of MERIS full resolution data to monitor vegetation seasonal dynamics. Remote Sens. Environ. 2009, 113, 1874–1885. [Google Scholar] [CrossRef]
  41. Wu, M.; Wu, C.; Huang, W.; Niu, Z.; Wang, C.; Li, W.; Hao, P. An improved high spatial and temporal data fusion approach for combining Landsat and MODIS data to generate daily synthetic Landsat imagery. Inf. Fus. 2016, 31, 14–25. [Google Scholar] [CrossRef]
  42. Dubrule, O. Comparing splines and kriging. Comput. Geosci. 1984, 10, 327–338. [Google Scholar]
  43. Chen, X.; Li, W.; Chen, J.; Rao, Y.; Yamaguchi, Y. A combination of TsHARP and thin plate spline interpolation for spatial sharpening of thermal imagery. Remote Sens. 2014, 6, 2845–2863. [Google Scholar] [CrossRef]
  44. Jarihani, A.; McVicar, T.; Van Niel, T.; Emelyanova, I.; Callow, J.; Johansen, K. Blending Landsat and MODIS data to generate multispectral indices: A comparison of “Index-then-Blend” and “Blend-then-Index” approaches. Remote Sens. 2014, 6, 9213–9238. [Google Scholar] [CrossRef]
  45. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: from error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef]
  46. Gorelick, N.; Hancher, M.; Dixon, M.; Ilyushchenko, S.; Thau, D.; Moore, R. Google Earth Engine: Planetary-scale geospatial analysis for everyone. Remote Sens. Environ. 2017, 202, 18–27. [Google Scholar] [CrossRef]
Figure 1. Flowchart of the proposed Enhanced Flexible Spatiotemporal DAta Fusion (EFSDAF) method.
Figure 1. Flowchart of the proposed Enhanced Flexible Spatiotemporal DAta Fusion (EFSDAF) method.
Applsci 09 03693 g001
Figure 2. Test data with gradual change (vegetation phenology) in Gwydir. (a) Landsat image of 5 July 2004; (b) Landsat image of 6 August 2004; (c) and (d) MODIS images of the same periods as Landsat images. All images use NIR-red-green as R-G-B.
Figure 2. Test data with gradual change (vegetation phenology) in Gwydir. (a) Landsat image of 5 July 2004; (b) Landsat image of 6 August 2004; (c) and (d) MODIS images of the same periods as Landsat images. All images use NIR-red-green as R-G-B.
Applsci 09 03693 g002
Figure 3. Test data with abrupt change (flood) in Gwydir. (a) Landsat image of 26 November 2004; (b) Landsat image of 12 December 2004; (c) and (d) MODIS images of the same periods to Landsat images. All images use NIR-red-green as R-G-B.
Figure 3. Test data with abrupt change (flood) in Gwydir. (a) Landsat image of 26 November 2004; (b) Landsat image of 12 December 2004; (c) and (d) MODIS images of the same periods to Landsat images. All images use NIR-red-green as R-G-B.
Applsci 09 03693 g003
Figure 4. Visual comparison of the predicted Landsat-like images by STARFM, STRUM, FSDAF, and EFSDAF in dataset 1 (vegetation phenology). From left to right, green-red-NIR (standard false color) composites of (a) a true Landsat image of 6 August 2004; (be) predicted images of 6 August 2004 by STARFM, STRUM, FSDAF, and EFSDAF, respectively. (Figure 4 shows the predicted Landsat-like images of 6 August 2004 by four fusion methods in gradual change area, and the yellow elliptical circles represent the intuitive differences of the four fusion methods. The zoom-in black square areas marked in Figure 4 are used to visually evaluate the performance of the four methods for the vegetation phenology in heterogeneous regions (Figure 5)).
Figure 4. Visual comparison of the predicted Landsat-like images by STARFM, STRUM, FSDAF, and EFSDAF in dataset 1 (vegetation phenology). From left to right, green-red-NIR (standard false color) composites of (a) a true Landsat image of 6 August 2004; (be) predicted images of 6 August 2004 by STARFM, STRUM, FSDAF, and EFSDAF, respectively. (Figure 4 shows the predicted Landsat-like images of 6 August 2004 by four fusion methods in gradual change area, and the yellow elliptical circles represent the intuitive differences of the four fusion methods. The zoom-in black square areas marked in Figure 4 are used to visually evaluate the performance of the four methods for the vegetation phenology in heterogeneous regions (Figure 5)).
Applsci 09 03693 g004
Figure 5. Zoomed in scenes of the black square areas marked in Figure 4. From left to right, (a) true Landsat image of 5 July 2004(base date); (b) true Landsat image of 6 August 2004(prediction date); (cf) predicted images of 6 August 2004 by STARFM, STRUM, FSDAF and EFSDAF, respectively.
Figure 5. Zoomed in scenes of the black square areas marked in Figure 4. From left to right, (a) true Landsat image of 5 July 2004(base date); (b) true Landsat image of 6 August 2004(prediction date); (cf) predicted images of 6 August 2004 by STARFM, STRUM, FSDAF and EFSDAF, respectively.
Applsci 09 03693 g005
Figure 6. Scatter plots of the true values and predicted values for NIR band in dataset 1. From left to right, (ad) the scatter plots for STARFM, STRUM, FSDAF and EFSDAF, respectively.
Figure 6. Scatter plots of the true values and predicted values for NIR band in dataset 1. From left to right, (ad) the scatter plots for STARFM, STRUM, FSDAF and EFSDAF, respectively.
Applsci 09 03693 g006
Figure 7. Visual comparison of Landsat-like images predicted by STARFM, STRUM, FSDAF, and EFSDAF in dataset 2 (flood). From left to right, green-red-NIR (standard false color) composites of (a) true Landsat image of 12 December 2004; (be) predicted images of 12 December 2004 by STARFM, STRUM, FSDAF, and EFSDAF, respectively. (Figure 7 presents the predicted Landsat-like images of 12 December 2004 by four fusion methods in the abrupt change area, and the yellow elliptical circles of Figure 7 represent the differences of the four fusion methods. The zoom-in black square areas marked in Figure 7 are used to more clearly compare the fusion results by the four methods in the land cover change area (Figure 8)).
Figure 7. Visual comparison of Landsat-like images predicted by STARFM, STRUM, FSDAF, and EFSDAF in dataset 2 (flood). From left to right, green-red-NIR (standard false color) composites of (a) true Landsat image of 12 December 2004; (be) predicted images of 12 December 2004 by STARFM, STRUM, FSDAF, and EFSDAF, respectively. (Figure 7 presents the predicted Landsat-like images of 12 December 2004 by four fusion methods in the abrupt change area, and the yellow elliptical circles of Figure 7 represent the differences of the four fusion methods. The zoom-in black square areas marked in Figure 7 are used to more clearly compare the fusion results by the four methods in the land cover change area (Figure 8)).
Applsci 09 03693 g007
Figure 8. Zoomed in scenes of the black square areas marked in Figure 7. From left to right, (a) true Landsat image of 26 November 2004 (base date); (b) true Landsat image of 12 December 2004 (prediction date); (cf) predicted images of 12 December 2004 by STARFM, STRUM, FSDAF, and EFSDAF, respectively.
Figure 8. Zoomed in scenes of the black square areas marked in Figure 7. From left to right, (a) true Landsat image of 26 November 2004 (base date); (b) true Landsat image of 12 December 2004 (prediction date); (cf) predicted images of 12 December 2004 by STARFM, STRUM, FSDAF, and EFSDAF, respectively.
Applsci 09 03693 g008
Figure 9. Scatter plots of the true values and the predicted values for NIR band in dataset 2. From left to right, (ad) the scatter plots for STARFM, STRUM, FSDAF and EFSDAF, respectively.
Figure 9. Scatter plots of the true values and the predicted values for NIR band in dataset 2. From left to right, (ad) the scatter plots for STARFM, STRUM, FSDAF and EFSDAF, respectively.
Applsci 09 03693 g009
Figure 10. Differences comparison between the predicted image by EFSDAF and input images in dataset 2. From left to right: (a) Landsat image at T P ; (b) Landsat-like image by EFSDAF at T P ; (c) MODIS image at T P ; (d) MODIS image at T B .
Figure 10. Differences comparison between the predicted image by EFSDAF and input images in dataset 2. From left to right: (a) Landsat image at T P ; (b) Landsat-like image by EFSDAF at T P ; (c) MODIS image at T P ; (d) MODIS image at T B .
Applsci 09 03693 g010
Table 1. The quantitative assessment of the Landsat-like images by STARFM, STRUM, FSDAF, and EFSDAF in dataset 1.
Table 1. The quantitative assessment of the Landsat-like images by STARFM, STRUM, FSDAF, and EFSDAF in dataset 1.
STARFMSTRUMFSDAFEFSDAF
ADB10.02670.02640.02620.0260
B20.02050.02030.01970.0197
B30.01760.01820.01690.0167
B40.04420.04800.03850.0379
B50.02440.02370.02330.0209
B70.02500.02660.02370.0226
Mean0.02640.02720.02470.0225
RMSEB10.03190.02990.02950.0284
B20.02700.02500.02450.0229
B30.02490.02440.02320.0212
B40.05880.06460.05420.0504
B50.03290.03220.03100.0270
B70.03410.03630.03220.0297
Mean0.03490.03540.03240.0299
CCB10.6530.7320.7780.803
B20.7240.7810.8120.845
B30.8520.8690.8950.915
B40.8910.8500.8970.914
B50.8560.8710.8800.918
B70.8650.8530.8860.905
Mean0.8070.8260.8580.883
SSIMB10.5040.5750.6170.632
B20.6830.7440.7980.804
B30.8360.8520.8650.890
B40.8850.8480.8930.902
B50.8540.8690.8870.908
B70.8620.8480.8810.892
Mean0.7710.7890.8230.839
Table 2. Quantitative assessment of the Landsat-like images by STARFM, STRUM, FSDAF, and EFSDAF in dataset 2.
Table 2. Quantitative assessment of the Landsat-like images by STARFM, STRUM, FSDAF, and EFSDAF in dataset 2.
STARFMSTRUMFSDAFEFSDAF
ADB10.01040.01040.00960.0094
B20.01370.01410.01300.0127
B30.01650.01670.01550.0151
B40.02490.02590.02380.0227
B50.04870.05200.04230.0425
B70.04440.04720.03940.0392
Mean0.02640.02770.02390.0235
RMSEB10.01520.01450.01370.0136
B20.02140.02070.01960.0193
B30.02600.02490.02360.0231
B40.03640.03710.03420.0326
B50.06430.06850.05600.0560
B70.05600.05950.05050.0502
Mean0.03660.03750.03290.0324
CCB10.5950.6170.6540.667
B20.6070.6280.6650.672
B30.5940.6260.6630.674
B40.7890.7710.8080.827
B50.7590.7250.7850.791
B70.7430.7100.7520.763
Mean0.6810.6790.7210.733
SSIMB10.5830.6110.6440.645
B20.5770.6130.6430.653
B30.5650.6100.6420.648
B40.7780.7610.7960.815
B50.7480.7140.7680.772
B70.7050.6690.7140.719
Mean0.6590.6630.7010.709
Table 3. The fusion precision of EFSDAF by using different endmember numbers in two datasets (NIR-red-green).
Table 3. The fusion precision of EFSDAF by using different endmember numbers in two datasets (NIR-red-green).
Dataset 1Dataset 2
EndmemberSVD345SVD345
ADGreen0.01970.01980.01970.01980.01270.01280.01300.0130
Red0.01670.01680.01680.01680.01510.01520.01560.0156
NIR0.03790.03900.03810.03740.02270.02270.02300.0236
RMSEGreen0.02290.02290.02290.02300.01930.01930.01970.0197
Red0.02120.02130.02130.02130.02310.02320.02360.0238
NIR0.05040.05040.04960.04900.03260.03270.03310.0336
CCGreen0.8450.8450.8460.8430.6720.6740.6560.657
Red0.9150.9140.9140.9140.6740.6750.6590.650
NIR0.9140.9140.9150.9180.8270.8270.8220.815
SSIMGreen0.8040.8040.8050.8020.6530.6450.6250.626
Red0.8900.8890.8890.8890.6480.6470.6350.618
NIR0.9020.9020.9060.9080.8150.8150.8100.800

Share and Cite

MDPI and ACS Style

Shi, C.; Wang, X.; Zhang, M.; Liang, X.; Niu, L.; Han, H.; Zhu, X. A Comprehensive and Automated Fusion Method: The Enhanced Flexible Spatiotemporal DAta Fusion Model for Monitoring Dynamic Changes of Land Surface. Appl. Sci. 2019, 9, 3693. https://doi.org/10.3390/app9183693

AMA Style

Shi C, Wang X, Zhang M, Liang X, Niu L, Han H, Zhu X. A Comprehensive and Automated Fusion Method: The Enhanced Flexible Spatiotemporal DAta Fusion Model for Monitoring Dynamic Changes of Land Surface. Applied Sciences. 2019; 9(18):3693. https://doi.org/10.3390/app9183693

Chicago/Turabian Style

Shi, Chenlie, Xuhong Wang, Meng Zhang, Xiujuan Liang, Linzhi Niu, Haiqing Han, and Xinming Zhu. 2019. "A Comprehensive and Automated Fusion Method: The Enhanced Flexible Spatiotemporal DAta Fusion Model for Monitoring Dynamic Changes of Land Surface" Applied Sciences 9, no. 18: 3693. https://doi.org/10.3390/app9183693

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop