Abstract
Cotton maps (10 m) of Xinjiang (XJ_COTTON10), which is the largest cotton production region of China, were produced from 2018 to 2021 through supervised classification. A two-step mapping strategy, i.e., cropland mapping followed by cotton extraction, was employed to improve the accuracy and efficiency of cotton mapping for a large region of about 1.66 million km2 with high heterogeneity. Additionally, the time-series satellite data related to spectral, textural, structural, and phenological features were combined and used in a supervised random forest classifier. The cotton/non-cotton classification model achieved overall accuracies of about 95% and 90% on the test samples of the same and adjacent years, respectively. The proposed two-step cotton mapping strategy proved promising and effective in producing multi-year and consistent cotton maps. XJ_COTTON10 agreed well with the statistical areas of cotton at the county level (R2 = 0.84–0.94). This is the first cotton mapping for the entire Xinjiang at 10-meter resolution, which can provide a basis for high-precision cotton monitoring and policymaking in China.
Similar content being viewed by others
Background & Summary
Cotton is one of the most important natural fibre and oil crops in the world1,2,3, and it can help to reduce extreme poverty (SDG 1, Sustainable Development Goal 1) and achieve the Zero Hunger Goal (SDG 2), as part of the 2030 Agenda for SDGs4. According to the Organization for Economic Co-operation and Development (OECD) and the Food and Agriculture Organization (FAO) of the United Nations, Xinjiang, as the leading and largest cotton production region in China, has accounted for over 80% and about 20% of cotton production in China and the world, respectively, during these years5,6. The accurate and detailed mapping of cotton in Xinjiang is vital for cotton management, disease prevention and control, and yield forecast. Xinjiang covers about one-sixth of China’s total land area, with an area of more than 1.66 million square kilometers (km2) and is characterized by typical heterogeneous desert-oasis ecotones (sourced from http://english.www.gov.cn/archive/ (last access: June 3, 2023)). It is a challenge to produce a fine cultivation map of cotton throughout Xinjiang7, probably because such a task needs a powerful platform with high storage and computing capabilities8; it needs enough high-quality and evenly distributed samples due to the high spatial heterogeneity9; and it needs models with high universality, accuracy, and transferability to address issues like bad weather10. To date, numerous studies have mainly focused on Xinjiang cotton mapping at county1,11,12, prefecture13, region7,14, and province scales15,16. Although we recently have produced the 500 m nationwide cotton maps in 2016 and 2018 that were included in the reports on remote sensing monitoring of China Sustainable Development15,16, both of them demonstrated limitations and disadvantages in practical precision management and decision making due to the very low spatial resolutions. To our knowledge, there were currently no fine-resolution large-scale cotton maps covering the entire Xinjiang.
It is of great importance to choose a suitable computing platform and remote sensing data for cotton mapping across Xinjiang, for satellite-based remote sensing imagery has been developed as the main data source for crop mapping in large spatial ranges17,18. With such data, a variety of crop mapping datasets have been produced and released successively, such as the Cropland Data Layer (CDL) of the United States19, the Agriculture and Agri-Food Canada’s Annual Crop Inventory (AAFC)20, the Early-Season Mapping of Winter Wheat in China (EMWC)21, and the Crop Mapping in Northeast China (CMNEC)22. Among them, the imagery sources mainly came from Landsat series21,23, SPOT series20, and Sentinel series22. Besides these large series datasets, GaoFen series and PlanetScope are another two datasets for the fine classification of crops24,25,26,27. However, the crop mapping for a vast area of a million km2 often needs to handle images with tens of thousands of scenes28, which are beyond the capacity of local workstations. The developing platforms with the capabilities of cloud storage and computing, such as the Google Earth Engine (GEE), have significantly aided the advancement of crop mapping at regional and global scales29,30. Ten years after the launch of GEE (2010–2019), more than 70 studies have been conducted in crop mapping using such a platform31.
Optical and radar remote sensing imagery are two main types of satellite data, both of which can provide abundant phenological and spatial information with spectral reflectance and backscattering coefficient, respectively18. Because of phenological differences among various crops, phenological characteristics and curves are often used to identify crops13,17,32. Spatial features such as the Gray-Level Co-occurrence Matrix (GLCM) indices have proven effective for improving the performance of crop classification33,34. As an active remote sensing technology, radar can effectively depict the spatial structure of land cover in cloudy and rainy weather, which is hard for optical remote sensing35. Due to their complementarity, the combination of optical and radar remote sensing has been a widely used strategy in crop classification1,7,28. In addition, there are two types of common auxiliary data, i.e., topographic and meteorological variables19,36. Topographic variables such as digital elevation model (DEM) and slope can aid in the identification of irrigated cropland37 and terraced fields38. Meteorological variables such as temperature and rainfall greatly affect plant growth and development and hence have strong correspondence with plant phenology39. Therefore, the combination of optical and radar remote sensing imageries as well as auxiliary data may offer more information gains for crop classification across Xinjiang.
In this study, we adopted all available Sentinel-1 and 2 imageries as well as auxiliary data on the GEE platform to develop 10 m cotton maps in Xinjiang from 2018 to 2021. Due to different availability of images in different years, we designed differentiated approaches and carried out detailed comparison and evaluation. The proposed method can be transferred to cotton mapping in a new year, and the published cotton maps dataset can provide a basis for crop management and policymaking.
Methods
Study area
Xinjiang is located in Northwest China (73°46′ ~ 96°23′E; 34°25′ ~ 49°50′N)40 (Fig. 1a,b), covering an area of about 1.66 million km2. It is the largest provincial administrative region in China, accounting for one-sixth of China’s total land area and with a land border of about 5600 kilometers (Fig. 1b). The terrain of Xinjiang is characterized by two basins (the Junggar Basin and Tarim Basin) between three mountains (Altay, Tianshan, and Kunlun Mountains) from north to south, and it is typically divided into three parts by the middle mountain Tianshan and climate zones, i.e., Northern Xinjiang (NX), Southern Xinjiang (SX), and Eastern Xinjiang (EX) (Fig. 1c,d). According to the Yearbooks of Xinjiang40,41, the main crops in Xinjiang are wheat, corn, and cotton, of which cotton is the largest crop in recent five years.
Data sets
According to the Yearbooks in Xinjiang42,43, we carried out field sampling in the main-cotton-planting areas in 2020 and 2021, respectively. A large number of positioning points for crops (including cotton, wheat, corn, and other crops) and non-crop (developed land, water, forest, grass, desert, mountain, etc.) were collected (Table 1). Further, to obtain enough high-quality and evenly distributed training data across the entire Xinjiang, sample expansion was performed by visual interpretation of very high-resolution imagery from the Google Earth platform7,13,14. As a result, a total of 10418 samples for 2020 (2627 samples for non-cotton crops and 7791 samples for non-crop) and a total of 5422 samples for 2021 (1147 samples for non-cotton crops and 4275 samples for non-crop) were acquired (Table 1). The spatial distributions of all samples in 2020 and 2021 are presented in Fig. 1e,f.
Surface reflectance
The Sentinel series imagery covering optical (Sentinel 2 A/B, S2) and radar (Sentinel 1, S1) imagery provides multi-band images with a spatial resolution of 10 m for Xinjiang from 2018 to 2021. S2 bottom-of-atmosphere (BOA) reflectance imagery from 2019 to 2021 and top-of-atmosphere (TOA) reflectance imagery from 2018 to 2021 are available in the GEE platform for Xinjiang. In general, S2 BOA reflectance imagery collection was often used to classify crops21,44,45. As shown in Fig. 1d,there are about 230 tiles in this study area, with a total of 31,945 scenes in 2018 (Table 2). Given the massive computing and storage demands, we did not pre-process the TOA images to BOA images. Instead, we used TOA imagery for cotton mapping in 2018 based on You et al.22 and You and Dong28.
To remove clouds and cloud shadows from TOA/BOA images, we adopted the following steps to process S2 data46:
-
(1)
obtaining clean pixels: employing the GEE-provided cloud probability dataset to generate clean pixels for each TOA/BOA image47,48.
-
(2)
Monthly synthesizing surface reflectance (SR): based on clean pixels, using the monthly median value as the SR value28.
-
(3)
Smoothing time-series data: eradicating null values and extreme values by deploying the Savitzky-Golay filter with a fitting window of 9 and a polynomial degree of 222.
Backscattering coefficient
Previous studies have reported that radar backscatter coefficients could capture crop structure, ground conditions, and their interaction information10,49. Therefore, S1 C-band (5.405 GHz) synthetic aperture radar (SAR) imagery collection was also included in this study (Table 3). The GEE platform provides the ground range detected (GRD) imagery with two bands of the interferometric wide (IW) swath mode, i.e., single co-polarization VV (vertical transmit/vertical receive) and dual-band cross-polarization VH (vertical transmit/horizontal receive). The provided imagery was pre-processed with the S1 Toolbox, sequentially by thermal noise removal, radiometric calibration, basic terrain correction for areas greater than 60° latitude, and finally scaling (10*log10(x)) as dB10,50. Besides, we further processed the GRD imagery with VV and VH bands by additional border noise correction, speckle filtering, and radiometric terrain normalization, according to Tian et al.10 and Mullissa et al.50 Additionally, similar to the S2 data, we performed denoising on the S1 data by the Savitzky-Golay filter, with a fitting window of 7 and a polynomial degree of 2.
Topographic variables
As shown in Fig. 1d, we employed the 30 m shuttle radar topography mission (SRTM) digital elevation model (DEM) data51 in this study. Besides DEM, the GEE platform provides its slope and aspect layers, which are often used in land cover change detection and classification37,38,52.
Meteorological variables
Previous studies have shown that temperature and rainfall exhibit strong correspondence with cropping patterns and phenological metrics in agricultural areas36,37,39. Annual mean temperature and precipitation from WorldClim Version 153 were employed here as two of the auxiliary datasets.
Geographical zones
In general, Xinjiang consists of NX, EX, and SX, as shown in Fig. 1c. These regions have different planting structures for major crops and different meteorological environments7,14,54. Referring to You et al.22, independent experiments and analyses were carried out in each geographical zone to eliminate spatial heterogeneity of cotton phenology.
Yearbooks
The statistical data in the yearbook is often used to evaluate the crop mapping performance21,22. There are two series of yearbooks in Xinjiang, i.e., the Yearbooks of Xinjiang Uygur Autonomous Region (hereinafter, Xinjiang(Local) for short) and the Yearbooks of Xinjiang Production and Construction Corps (hereinafter, Xinjiang(Corps) for short). They are independently compiled by the Statistic Bureaus of Xinjiang(Local) and Xinjiang(Corps), respectively. There are a total of 94 counties (including county-level cities) from Xinjiang(Local) (85 counties) and Xinjiang(Corps) (9 counties). Up to now, the annual planting areas from 2018 to 2020 have been published through the two series of yearbooks40,41,43,55,56.
Feature selection
In order to avoid overfitting and improve generalization ability, we preferred multi-sensor features with high independence considering spectral, textural and structural properties, respectively. We hence chose EVI (enhanced vegetation index, Eq. (1)) from S222,28, SAVG (sum average, Eq. (2)) of the blue band from S234,57, and the VH band from S123,28 as the main independent variables.
where p(i,j) is the (i,j)th entry in the GLCM of a pixel; Ng is the number of distinct gray levels in the quantized image58. In each GLCM, the size of the neighborhood was set to 4, and the kernel was set to a 9 × 9 square (radius = 4 pixels).
Besides, topographic and meteorological factors are important in distinguishing croplands37,38,52. Therefore, we employed three topographic variables and two meteorological variables, i.e., DEM, Slope, Aspect, temperature (Temp), and precipitation (Prec).
In terms of phenological curves between SX and NX, different crops and non-crop presented varied differences, according to Fig. 2. First, the EVI curves presented more obvious differences between crops and non-crop in NX than in SX. The maximum EVI values of wheat appeared in different months in SX (May) and NX (June). The EVI values of cotton from June to September were higher in NX than in SX. Second, the SAVG feature from December to February expressed obvious differences between crops and non-crop in NX but not in SX. Third, the VH curves indicated the stronger separability between cotton and non-cotton crops in NX than in SX. Moreover, the auxiliary features revealed that crops and non-crop were more separable in NX than in SX, especially Slope, Temp, and Prec. These details hence suggested the necessity of the zoning mapping strategy.
According to recent studies in crop classification and mapping7,10,21,22, by employing a two-step strategy, we planned to distinguish cropland from other lands and then performed the cotton/non-cotton classification within the cropland to map cotton. The two-step strategy can eliminate interference from non-cropland types and thus make cotton mapping more convenient and accurate. As seen in Fig. 2, some features, such as EVI and VH (from June to September), had great potential to distinguish cotton and non-cotton types, but may misclassify wheat fields and non-crop lands. By contrast, the EVI features from April to June had high potential for distinguishing cropland from non-crop lands but can hardly be used to classify crops, especially in NX. These preliminary results revealed the importance of the two-step mapping approach and the necessity of a classifier that can effectively select suitable features from the feature set.
Besides, there were obvious similarities in outstanding features in distinguishing cropland and cotton. For example, the SAVG feature in May can distinguish cotton from non-cotton and non-crop lands in both SX and NX. Notably, the feature curves derived from TOA imagery were quite similar to those derived from BOA imagery, indicating that TOA imagery can be used for crop mapping in this study.
Cotton mapping strategy
Given that a classification model in one spacetime can be transferred to another similar one59 and that crops in adjacent years and regions have similar characteristics, including phenological, topographic, and meteorological variables, transfer learning was often used in crop classification in Xinjiang11,12,15,16. As illustrated in Fig. 1e,f and Table 1, there were enough samples in 2020 and 2021 but none in 2018–2019. Meanwhile, we did not perform field sampling in EX because it is not a main-cotton-planting area. Therefore, transfer learning was employed: (1) transferring the models trained on the samples in 2020 to distinguish cropland and cotton in 2018–2019; and (2) transferring the models trained on the samples in NX to distinguish cropland and cotton in EX.
As shown in Table 2, the GEE platform provides no BOA imagery for this study area in 2018. Therefore, we designed two combinations about the selected features: (1) BOA+: time-series EVI and SAVG derived from the S2 BOA imagery, time-series VH, and the auxiliary features; and (2) TOA+: time-series EVI and SAVG derived from the S2 TOA imagery, time-series VH, and the auxiliary features.
The GEE platform provides multiple machine learning methods for imagery classification, such as random forest (RF), decision tree, support vector machine, and naive bayes classifiers. Among them, RF is commonly used in crop classification22,23,28,38. Because the RF classifier can employ the bagging technique to seek optimal features via the auto optimization mechanism for better classification60, we performed the models’ training with the same feature set for the cropland mapping and the cotton mapping. Taking the second step of cotton mapping as an example, the technical framework was provided in Fig. 3. It mainly included the training and test of the classification model in 2020 (Evaluation in 2020), the performance evaluation of the transfer ability in 2021 (Evaluation in 2021), and cotton planting area assessment in 2018–2020 (Assessment in 2018/2019/2020). Finally, we can obtain the 10-m cotton maps during 2018–2021 via the TOA+ approach and three 10-m cotton maps during 2019–2021 via the BOA+ approach.
For evaluating the models in 2020 and 2021, we adopted the following five metrics: OA (overall accuracy), Kappa (Kappa coefficient), UA (user’s accuracy), PA (producer’s accuracy), and the F1-score28. On the assessment of cotton planting areas between the estimation and the summary, we employed R2 (the coefficient of determination), p value (the significance level), and RMSE (the root mean square error)21.
Postprocessing
The salt and pepper noise often appears in the classification result of a vast area, mainly because of different objects with the same spectrum or the same objects with different spectra38,61. To address this issue, edge-preserving filtering with a bilateral filter62 was employed. For the used bilateral filter, the standard deviation of the spatial Gaussian smoothing kernel was set to 1, and the size of the square neighborhood around each pixel was set to 5 pixels63.
Data Records
The XJ_COTTON10 dataset64 (10-m cotton maps of Xinjiang from 2018 to 2021) is available to the public at https://doi.org/10.5281/zenodo.7856467. It includes the following seven compressed packages of 30 geofiles with a size of 3° × 3°:
-
1)
Cotton_2018_TOA_10m_filter.zip: the 10 m cotton map in 2018 via the TOA+ approach;
-
2)
Cotton_2019_BOA_10m_filter.zip: the 10 m cotton map in 2019 via the BOA+ approach;
-
3)
Cotton_2019_TOA_10m_filter.zip: the 10 m cotton map in 2019 via the TOA+ approach;
-
4)
Cotton_2020_BOA_10m_filter.zip: the 10 m cotton map in 2020 via the BOA+ approach;
-
5)
Cotton_2020_TOA_10m_filter.zip: the 10 m cotton map in 2020 via the TOA+ approach;
-
6)
Cotton_2021_BOA_10m_filter.zip: the 10 m cotton map in 2021 via the BOA+ approach;
-
7)
Cotton_2021_TOA_10m_filter.zip: the 10 m cotton map in 2021 via the TOA+ approach.
The geographic reference (ESPG: 4326 (WGS_1984)) is the same for all seven maps, conforming to that in the geotiff file. The values of the seven maps contains 1 and 0, representing cotton and other land (including other crops and non-cropland).
Technical Validation
Model evaluation
During the RF classifier training and validation, the randomly selected half of the total samples for each type were used as the training samples, and the rest were used as the test samples, to ensure enough samples in both training and test22,65. The number of trees was set to 100, and the number of variables per split was set as the square root of the number of variables (41)22. The estimated labels of the test samples can be achieved by a corresponding classification model for each zone (Fig. 1c). We combined the estimated labels and ground truth labels of all zones together to produce a total confusion matrix. Table 4 illustrates the summary of the accuracy assessment of crop/non-crop classification models on samples in 2020 and 2021. The crop mapping models of the two strategies (testing by the samples in the same year and the adjacent year) had OAs of over 95% and F1-scores of cropland of over 95%, indicating that transfer learning was feasible on crop mapping across Xinjiang. Besides, the two feature-combinations of TOA+ and BOA+ had close performance on the five statistical indices, suggesting that S2 TOA imagery worked well on cropland mapping in this study area.
According to the summary of the accuracy assessment of cotton/non-cotton classification models on samples in 2020 and 2021 (Table 5), The strategy (M20 and M21) of testing by the samples in the same year achieved OAs of over 94% and F1-scores of cotton of over 95%, which were about 4–5% and about 3% higher than the transfer learning strategy M20T21, respectively. Despite all this, the transfer learning strategy, which had an OA of over 90% and an F1-score of cotton of over 92%, can be considered for cotton mapping in adjacent years without samples. Moreover, S2 TOA imagery also worked well on cotton mapping, considering the small performance differences between the two feature-combinations of TOA+ and BOA+.
Area assessment
Using the cotton mapping strategies, cotton maps during 2018–2021 via the TOA+ and BOA+ approaches were produced on the GEE platform. Notably, cropland and cotton maps in 2018, 2019, and 2021 were developed by the transfer learning strategy of the models trained on the samples in 2020. We further computed the total cotton-planting area at the county level for the seven cotton maps on the GEE platform. Combining the estimated areas by cotton maps and the statistical areas by the two series of yearbooks in Xinjiang, the performance evaluation of the cotton mapping was delivered in Fig. 4. Results showed that our maps accurately estimated the areas of cotton at the county level, with R2 values ranging from 0.84 to 0.94. By contrast, the transfer learning strategy based on BOA+ (R2 = 90 in 2019) provided slightly better performance than TOA+ (R2 = 0.84–0.85 in 2019 and 2018). Moreover, from the comparisons between TOA+ and BOA+ (Fig. 4f–h), there were slight differences in county-level cotton-planting area estimations between them, suggesting the feasibility of the TOA+ strategy.
Uncertainty analysis
We adopted the classification probability of each pixel in cropland maps (cotton/non-cotton) to generate its uncertainty by 1-Pmax, where Pmax was the maximum probability of being classified as cotton/non-cotton types38,66. As seen in Fig. 5, by contrast, the uncertainties of crop pixels were lower in NX than in SX overall. Cropland fragmentation may be a main reason. In NX, there is a large amount of farmland that has been integrated into contiguous areas by Xinjiang(Corps) since the 1950s67,68. However, in SX, the majority of croplands were managed by individual farmers, resulting in fragmented lands with diverse crops67. As shown in Fig. 6, the cotton proportion maps at the kilometer scale revealed that there were large areas of concentrated and contiguous high-standard farmland for cotton planting in NX but large areas of dispersed cotton in SX. Previous studies have indicated that cropland fragmentation and heterogenization can make farmland management inefficient69 and crop yields decrease70,71,72. Such a land use pattern can lead to a large number of mixed pixels and decrease the representativeness of field samples, leading to uncertainty in classification73.
According to cotton distribution during 2018–2021, cotton is mainly in SX and NX (Fig. 6). At the prefecture level, the most densely distributed areas were in Bortala, Tarbagatay, Changji, Aksu, and Bayingolin. By contrast, cotton-planting fields were dispersed overall in SX but more concentrated in NX. We selected seven typical cotton-planting cases, which are in Alaer (Aksu), Tumushuke (Kashgar), Yizhou (Hami), Shawan (Tarbagatay), Wusu (Tarbagatay), Manas (Changji), and Bole (Bortala), and detailed their cotton dynamics during the last four years (Fig. 7). Visually, our cotton maps realistically provided most of the distributions and the dynamics of cotton. The blue boxes marked in Fig. 7b–g displayed a few cotton parcels with typical dynamic changes, and our maps clearly captured the details. For example, the displayed cotton-planting frequency in the blue boxes in Fig. 7e was in line with the dynamic of the field in these four years. The dynamics of cotton like these were generally accurately captured in our maps. Apart from the misclassified pixels near the edge of field parcels due to the mixed pixel issue, a number of roads were clearly seen between the parcels, highlighting the advantage of 10-m field maps over coarse maps15,16.
Transfer learning in crop classification is based on the basic postulate that crops in adjacent years and regions have similar characteristics including phenological, topographic and meteorological variables. However, as the interval increases, the similarity of the same crop in different years may lessen. Overall, the uncertainties in 2020 were lower than those in the other three years (Fig. 5).
To tackle the two problems mentioned above, adding more independent samples with high representativeness, especially in SX, and estimating cotton distributions at the subpixel level are two routes worth considering in future work.
Feature importance evaluation
Clear acquaintance with the key features in crop/cotton mapping is helpful for future research. There are two common ways to derive the relative importance scores of features used in RF models. One way is based on Gini index38, and the other one is based on the out-of-bag (OOB) error74,75. Figure 8 shows the OOB error-based feature importance values (FIVs) of different classification models in this study. There were lots of discoveries in Fig. 8 which were highly consistent with Fig. 2. For example, the SAVG feature in May (SAVG-5) was one of the main features in the cotton/non-cotton classification and was more important in NX than in SX. The Aspect feature was the worst in both the crop/non-crop and cotton/non-cotton classifications, and the Slope feature was more important in NX than in SX in the crop/non-crop classification.
By contrast, FIVs of VH (Sentinel-1) were 0.75 ± 0.20 in cropland mapping and 0.45 ± 0.20 in cotton mapping; FIVs of EVI (Sentinel-2) were 0.78 ± 0.18 in cropland mapping and 0.49 ± 0.14 in cotton mapping; and FIVs of SAVG (Sentinel-2) were 0.69 ± 0.17 in cropland mapping and 0.65 ± 0.36 in cotton mapping. Results showed that the SAVG feature of Sentinel-2 contributed more than EVI (Sentinel-2) and VH (Sentinel-1) in cotton mapping (cotton/non-cotton classification). While EVI (Sentinel-2) and VH (Sentinel-1) had similar performance in cropland mapping (crop/non-crop classification), both of which had significant advantages over SAVG.
To offer more convincing explanations for the selected features, we added three additional experiments: using (1) each of vegetation indices (VIs) (EVI, NDVI (normalized difference vegetation index), NIRv (near-infrared reflectance of vegetation), and NDWI (normalized difference water index)); (2) each of SAR bands (VV and VH); and (3) each of EVI-VH (the combination of EVI and VH) and EVI-SAVG (the combination of EVI and SAVG(Blue)) as the explanatory variables to classify cotton/non-cotton using the RF classifier in the three strategies of M20, M21, and M20T21 (Fig. S1). Results showed that, first, among the optical VIs, EVI achieved higher performance than the other three VIs, particularly in the strategy of M20T21; among the SAR bands, VH achieved higher performance than VV. And thus, we chose EVI and VH as two main explanatory variables for cotton/non-cotton classification. Second, the combination of EVI-VH achieved higher and more robust accuracies than EVI and VH in terms of the three strategies of M20, M21, and M20T21 (Table S1). In particular, EVI-VH derived from Sentinel 1 and 2 achieved higher and more robust performance than EVI-SAVG considering Sentinel-2 only.
Besides, there were new findings from Fig. 8. The optical and radar features (EVI, Blue-SAVG, and VH) in the fallow period (from November to February) can make contributions to the crop/non-crop classification but gave small gains for the cotton/non-cotton classification. The reason may include that there could be apparent textural and structural differences between crop parcels and non-crop (developed land, water, desert, mountain, etc.) during the fallow periods of crops. However, it may be hard to distinguish different croplands during their fallow periods from the perspectives of texture and structure. Therefore, these variables in the fallow period can be eliminated in the cotton/non-cotton classification in the future. As the two long-term indicators, annual mean temperature (Temp) and precipitation (Prec) contributed significantly to identifying both cropland and cotton. All the three time-series features provided relatively balanced contributions, indicating their complementary advantages.
Summary
Our 10-m cotton maps from 2018 to 2021 were derived from our two-step classification models with high and satisfactory accuracy and exhibited good visualization and realistic dynamics of cotton across Xinjiang. In this study, we employed the monthly interval to generate time-series features, and such an interval was proven to be feasible in weakening intra-class differences and improving inter-class separability (Fig. 2). Besides, You et al.22 reported that the monthly interval can meet the demand that almost all locations have at least one clear observation (after cloud and cloud shadow mask) in each period, but other intervals (i.e., 10-, 15-, and 20-day) cannot. Besides, the monthly composite may further weaken the influence of cotton phenological differences within the same geographical zone (Fig. 1c)76, thereby yielding cotton/non-cotton classification models with higher performance. In addition, the monthly composite-based classification models could be more concise and require lower cloud computing capabilities than other composite ways (i.e., 10-, 15-, and 20-day). The Sentinel 1B satellite has not worked since December 23rd, 2021, according to the news about the Sentinel-1 mission (https://sentinels.copernicus.eu/web/sentinel/missions/sentinel-1 (last access: June 3, 2023)). Despite the fact that Sentinel-1A was already nearing its maximum capacity, the number of effective observations per unit time of S1 was still significantly reduced, as shown in Fig. 9a. At present, enough observations (at least once a month) can still be provided by the S1 and S2 products (Fig. 9). Therefore, the proposed two-step strategy for cotton mapping can be directly transferred to 2022 in future work. However, numerous limitations and deficiencies are worth considering in future research.
First, besides the four 10-m bands (B2, B3, B4, and B8), there are other eight bands with relatively coarse resolutions (i.e., 20-m and 60-m) in S2 imagery. In order to produce the 10-m cotton maps, we only take the bands with the 10-m resolution as the basic variables. However, except for the 30-m DEM and the about 900-m WorldClim products, the GEE platform cannot provide more refined auxiliary datasets. The inconsistent resolutions of the basic and auxiliary variables probably increased the uncertainty. Besides, our classification models adopted features from multiple sensors, considering the strong independencies among different sensors. On the one hand, multiple vegetation indices and red-edge bands are often employed in the mapping of major crops22,28,44, indicating they may be suitable for cotton mapping in Xinjiang, but on the other hand, multiple time-series vegetation indices and bands as independent variables may lead to the Hughes phenomenon77. Therefore, choosing how many and which vegetation indexes as explanatory variables for cotton mapping is worth attempting in the future.
Second, earliest identifiable timing (EIT) and early-season mapping of crops are highly important for crop monitoring and yield forecasting21,28,78. The harvest stage of cotton in Xinjiang is mainly from October to November13,79. As illustrated in Fig. 2, the strong separability between cotton and non-cotton crops can be given by EVI from June to September, SAVG from April to June, and VH from June to September. And thus, the selected three features had the potential for EIT and early-season mapping of cotton in this study.
Third, some typical phenological features in the flowering and boll-opening stages may have potential for cotton mapping. However, it is hard to recognize the cotton flower by the optical imagery of Sentinel 2, mainly because there are more than ten flower colors for cotton, including white and red colors that have obvious color differences80,81. Particularly, (1) petal colors could change during flower development80; (2) there are also significant differences in color changes for different cotton varieties80; and (3) even on the same cotton plant, flowers of different colors coexist82. The boll-opening stage is another important phenological period. Most cotton has white fibre, and a small amount of cotton has colored fibre in Xinjiang83,84. A few recent studies have reported that the phenological feature in the boll-opening stage could be used for cotton mapping11 and cotton yield estimation85,86,87,88. Wang et al. (2021) proposed a white bolls index (WBI) using Sentinel 2 imagery in the boll-opening stage and performed cotton mapping in a county in Xinjiang11. WBI achieved the mapping performance (OA = 95% and Kappa = 0.88) that is similar to ours in this study, but it may have the following deficiencies: first, it needs the time-series data at a higher temporal resolution (5-day) to determine the time of boll-opening. In other words, it does not reduce the required data, in comparison with the monthly vegetation index used in our work. Second, it relies heavily on the optical imagery in the boll-opening period. However, it is difficult to guarantee that all cotton parcels have at least one clear observation during the boll-opening stage throughout Xinjiang due to the influence of the bad weather (clouds, mists, and rain). Third, it is hard for the approach of WBI to be transferred to the earliest identification of cotton. In particular, WBI is used to distinguish cotton and non-cotton, and hence it is essential to obtain the cropland using another technical method.
In addition, the classification models need to be improved further because the uncertainties increased when the models were transferred to adjacent years. Up to now, more and more case studies indicate that deep learning may have stronger generalization capability and transferability in crop classification than traditional machine learning89,90,91. However, deep learning models are performed on high-performance computers, and the present public platforms, including GEE, do not provide the service for deep learning. If only a deep learning strategy is used for 10-m cotton mapping in Xinjiang, it will be very expensive. The cropland area accounts for about 3% of the total land area in Xinjiang40,55. A compromise scheme, i.e., cropland mapping by traditional machine learning and then cotton mapping by deep learning, may be promising.
To make comparisons between the used RF models and deep learning-based models, we added an additional experiment of a spatio-temporal data-driven deep learning architecture named CNN-LSTM (a hybrid network combining convolutional neural network (CNN) and long short-term memory (LSTM))92,93 for the cotton/non-cotton classification (Fig. S2): first, we pre-processed and downloaded the surface reflectance data of all samples and their surrounding pixels (monthly synthesis) from GEE; second, similar to the model evaluation of RF classification performed on the GEE platform, we adopted the three strategies M20, M21, and M20T21 for the CNN-LSTM model in MATLAB 2021b (The MathWorks, Inc., Natick, MA, USA) installed on a personal computer. As shown in Fig. S3, the CNN-LSTM model achieved competitive performance in cotton/non-cotton classification for M20 and M21 but relative low performance for M20T21 under different combinations of bands (CBs) and patch sizes, compared with the proposed RF model based on multi-source data (Table 5). For M20T21, the best CNN-LSTM model was found when CBs = 9 and patch size = 5. It had an OA of 85.83%, a Kappa of 0.69, and an F1-score of cotton of 88.93%. This model was significantly worse than the proposed RF model based on multi-source data (Table 5). Considering that cotton maps in 2018, 2019, and 2021 were created using the model trained on the samples in 2020, we would prefer to choose the RF classification model instead of CNN-LSTM. Further research into the new architectures, such as the network structure considering the spatiotemporal attention mechanism, may be promising in improving the interannual transferability94,95.
Usage Notes
Xinjiang is the leading and largest cotton production region in China and has accounted for over 80% and about 20% of cotton production in the whole nation and the world, respectively5,6. In this study, we provided the XJ_COTTON10 product64 (Fig. 6) from 2018 to 2021 through supervised classification using high-quality in-field samples and multi-source remote sensing data on the GEE platform. Compared with the latest published cotton mapping products15,16, the cotton maps of the four years in this study exhibited better spatial details and more realistic dynamics of cotton across Xinjiang (Fig. 7). In comparison with the two series of yearbooks in Xinjiang, the cotton maps gave the estimated areas with high accuracy of cotton at the county level. These novel 10-m cotton maps, which explained an average of 89.7% of the spatial variabilities in the planting areas at the county level during 2018–2020, could be helpful for high-precision cotton monitoring and policymaking across the whole province and cotton yield and production forecasting at the field parcel scale (Fig. 7).
Code availability
The GEE codes used to generate the cropland and cotton maps are available to the public at https://zenodo.org/record/785646764.
References
Xun, L., Zhang, J., Cao, D., Yang, S. & Yao, F. A novel cotton mapping index combining Sentinel-1 SAR and Sentinel-2 multispectral imagery. ISPRS Journal of Photogrammetry and Remote Sensing 181, 148–166, https://doi.org/10.1016/j.isprsjprs.2021.08.021 (2021).
Huang, J. K., Hu, R. F., van Meijl, H. & van Tongeren, F. Biotechnology boosts to crop productivity in China: trade and welfare implications. J. Dev. Econ. 75, 27–54, https://doi.org/10.1016/j.jdeveco.2003.07.005 (2004).
Riaz, T. et al. Cottonseed oil: A review of extraction techniques, physicochemical, functional, and nutritional properties. Critical Reviews in Food Science and Nutrition, 1–19 (2021).
Amrouk, E. M., Mermigkas, G. & Townsend, T. Recent trends and prospects in the world cotton market and policy developments. (Food and Agriculture Organization of the United Nations FAO, 2021).
OECD & FAO. OECD-FAO Agricultural Outlook 2021-2030. (OECD Publishing, 2021).
Fu, L. & Liu, A. China Statistical Yearbook - 2021. (China Statistics Press, 2021).
Hu, T., Hu, Y., Dong, J., Qiu, S. & Peng, J. Integrating Sentinel-1/2 Data and Machine Learning to Map Cotton Fields in Northern Xinjiang, China. Remote Sensing 13, https://doi.org/10.3390/rs13234819 (2021).
Yang, L. et al. Google Earth Engine and Artificial Intelligence (AI): A Comprehensive Review. Remote Sensing 14, https://doi.org/10.3390/rs14143253 (2022).
Lang, P. et al. Integrating environmental and satellite data to estimate county-level cotton yield in Xinjiang Province. Frontiers in Plant Science 13, https://doi.org/10.3389/fpls.2022.1048479 (2023).
Yang, G. et al. Numerical assessment of the effect of water-saving irrigation on the water cycle at the Manas River Basin oasis, China. Sci. Total Environ. 707, 8, https://doi.org/10.1016/j.scitotenv.2019.135587 (2020).
Wang, N., Zhai, Y. & Zhang, L. Automatic cotton mapping using time series of Sentinel-2 images. Remote Sensing 13, 1355 (2021).
Hao, P., Wang, L., Zhan, Y. & Niu, Z. Using moderate-resolution temporal NDVI profiles for high-resolution crop mapping in years of absent ground reference data: a case study of bole and manas counties in Xinjiang, China. ISPRS International Journal of Geo-Information 5, 67 (2016).
Liu, G. Understanding cotton cultivation dynamics in Aksu Oases (NW China) by reconstructing change trajectories using multi-temporal Landsat and Sentinel-2 data. Geocarto International, 1–19, https://doi.org/10.1080/10106049.2021.1886337 (2021).
Liu, C. et al. Analysis of the spatial distribution and variation characteristics of cotton planting in southern Xinjiang from 2000 to 2020. Transactions of the Chinese Society of Agricultural Engineering (Transactions of the CSAE) 37, 223–232, https://doi.org/10.11975/j.issn.1002-6819.2021.16.028 (2021).
Gu, X. et al. Report on Remote Sensing Monitoring of China Sustainable Development (2017). (Social Sciences Academic Press (CHINA), 2018).
Gu, X. et al. Report on Remote Sensing Monitoring of China Sustainable Development (2019). (Social Sciences Academic Press (CHINA), 2020).
Dong, J. W. & Xiao, X. M. Evolution of regional to global paddy rice mapping methods: A review. Isprs Journal of Photogrammetry and Remote Sensing 119, 214–227, https://doi.org/10.1016/j.isprsjprs.2016.05.010 (2016).
Orynbaikyzy, A., Gessner, U. & Conrad, C. Crop type classification using a combination of optical and radar remote sensing data: a review. International Journal of Remote Sensing 40, 6553–6595, https://doi.org/10.1080/01431161.2019.1569791 (2019).
Boryan, C., Yang, Z., Mueller, R. & Craig, M. Monitoring US agriculture: the US department of agriculture, national agricultural statistics service, cropland data layer program. Geocarto International 26, 341–358 (2011).
Fisette, T. et al. in 2013 Second International Conference on Agro-Geoinformatics (Agro-Geoinformatics). 270–274 (IEEE).
Dong, J. et al. Early-season mapping of winter wheat in China based on Landsat and Sentinel images. Earth System Science Data 12, 3081–3095, https://doi.org/10.5194/essd-12-3081-2020 (2020).
You, N. et al. The 10-m crop type maps in Northeast China during 2017-2019. Sci Data 8, 41, https://doi.org/10.1038/s41597-021-00827-9 (2021).
Demarez, V., Helen, F., Marais-Sicre, C. & Baup, F. In-Season Mapping of Irrigated Crops Using Landsat 8 and Sentinel-1 Time Series. Remote Sensing 11, https://doi.org/10.3390/rs11020118 (2019).
Song, Q. et al. Mapping regional cropping patterns by using GF-1 WFV sensor data. Journal of Integrative Agriculture 16, 337–347, https://doi.org/10.1016/s2095-3119(16)61392-8 (2017).
Song, Q. et al. In-Season Crop Mapping with GF-1/WFV Data by Combining Object-Based Image Analysis and Random Forest. Remote Sensing 9, 19, https://doi.org/10.3390/rs9111184 (2017).
Rao, P. et al. Using Sentinel-1, Sentinel-2, and Planet Imagery to Map Crop Type of Smallholder Farms. Remote Sensing 13, 13, https://doi.org/10.3390/rs13101870 (2021).
Pickering, J. et al. Using Multi-Resolution Satellite Data to Quantify Land Dynamics: Applications of PlanetScope Imagery for Cropland and Tree-Cover Loss Area Estimation. Remote Sensing 13, 15, https://doi.org/10.3390/rs13112191 (2021).
You, N. & Dong, J. Examining earliest identifiable timing of crops using all available Sentinel 1/2 imagery and Google Earth Engine. ISPRS Journal of Photogrammetry and Remote Sensing 161, 109–123, https://doi.org/10.1016/j.isprsjprs.2020.01.001 (2020).
Teluguntla, P. et al. A 30-m landsat-derived cropland extent product of Australia and China using random forest machine learning algorithm on Google Earth Engine cloud computing platform. Isprs Journal of Photogrammetry and Remote Sensing 144, 325–340, https://doi.org/10.1016/j.isprsjprs.2018.07.017 (2018).
Xiong, J. et al. Automated cropland mapping of continental Africa using Google Earth Engine cloud computing. Isprs Journal of Photogrammetry and Remote Sensing 126, 225–244, https://doi.org/10.1016/j.isprsjprs.2017.01.019 (2017).
Tamiminia, H. et al. Google Earth Engine for geo-big data applications: A meta-analysis and systematic review. Isprs Journal of Photogrammetry and Remote Sensing 164, 152–170, https://doi.org/10.1016/j.isprsjprs.2020.04.001 (2020).
Zhang, L. et al. Crop classification based on the spectrotemporal signature derived from vegetation indices and accumulated temperature. International Journal of Digital Earth, 1-27 (2022).
Peña-Barragán, J. M., Ngugi, M. K., Plant, R. E. & Six, J. Object-based crop identification using multiple vegetation indices, textural features and crop phenology. Remote Sensing of Environment 115, 1301–1316, https://doi.org/10.1016/j.rse.2011.01.009 (2011).
Tassi, A. & Vizzari, M. Object-Oriented LULC Classification in Google Earth Engine Combining SNIC, GLCM, and Machine Learning Algorithms. Remote Sensing 12, https://doi.org/10.3390/rs12223776 (2020).
Shao, Y. et al. Rice monitoring and production estimation using multitemporal RADARSAT. Remote Sensing of Environment 76, 310–325, https://doi.org/10.1016/s0034-4257(00)00212-1 (2001).
Arévalo, P., Bullock, E. L., Woodcock, C. E. & Olofsson, P. A suite of tools for continuous land change monitoring in google earth engine. Frontiers in Climate, 26 (2020).
Thenkabail, P. S. et al. Global irrigated area map (GIAM), derived from remote sensing, for the end of the last millennium. International Journal of Remote Sensing 30, 3679–3733, https://doi.org/10.1080/01431160802698919 (2009).
Cao, B. W. et al. A 30 m terrace mapping in China using Landsat 8 imagery and digital elevation model based on the Google Earth Engine. Earth System Science Data 13, 2437–2456, https://doi.org/10.5194/essd-13-2437-2021 (2021).
Zhang, X. Y., Friedl, M. A. & Schaaf, C. B. Global vegetation phenology from Moderate Resolution Imaging Spectroradiometer (MODIS): Evaluation of global patterns and comparison with in situ measurements. J. Geophys. Res.-Biogeosci. 111, 14, https://doi.org/10.1029/2006jg000217 (2006).
Gao, W. & Han, R. Xinjiang Statistical Yearbook - 2021. (China Statistics Press, 2021).
Liu, Y. et al. Resistance and resilience of grasslands to drought detected by SIF in inner Mongolia, China. Agricultural and Forest Meteorology 308-309, https://doi.org/10.1016/j.agrformet.2021.108567 (2021).
Gao, W. & Han, R. Xinjiang Statistical Yearbook - 2019. (China Statistics Press, 2019).
Li, W. & Yu, X. Xinjiang Production & Construction Corps Statistical Yearbook - 2019. (China Statistics Press, 2019).
Yi, Z., Jia, L. & Chen, Q. Crop Classification Using Multi-Temporal Sentinel-2 Data in the Shiyang River Basin of China. Remote Sensing 12, https://doi.org/10.3390/rs12244052 (2020).
Lozano-Tello, A. et al. Crop identification by massive processing of multiannual satellite imagery for EU common agriculture policy subsidy control. European Journal of Remote Sensing 54, 1–12, https://doi.org/10.1080/22797254.2020.1858723 (2021).
Kang, X., Huang, C., Zhang, L., Zhang, Z. & Lv, X. Downscaling solar-induced chlorophyll fluorescence for field-scale cotton yield estimation by a two-step convolutional neural network. Computers and Electronics in Agriculture 201, 107260, https://doi.org/10.1016/j.compag.2022.107260 (2022).
Xu, L. et al. Time series analysis for global land cover change monitoring: A comparison across sensors. Remote Sensing of Environment 271, https://doi.org/10.1016/j.rse.2022.112905 (2022).
Pazúr, R., Huber, N., Weber, D., Ginzler, C. & Price, B. A national extent map of cropland and grassland for Switzerland based on Sentinel-2 data. Earth System Science Data 14, 295–305 (2022).
Veloso, A. et al. Understanding the temporal behavior of crops using Sentinel-1 and Sentinel-2-like data for agricultural applications. Remote Sensing of Environment 199, 415–426, https://doi.org/10.1016/j.rse.2017.07.015 (2017).
Mullissa, A. et al. Sentinel-1 SAR Backscatter Analysis Ready Data Preparation in Google Earth Engine. Remote Sensing 13, https://doi.org/10.3390/rs13101954 (2021).
Farr, T. G. et al. The Shuttle Radar Topography Mission. Reviews of Geophysics 45, https://doi.org/10.1029/2005rg000183 (2007).
Boryan, C., Yang, Z., Willis, P. & Sandborn, A. in IEEE International Geoscience and Remote Sensing Symposium 6239–6242 (2019).
Hijmans, R. J., Cameron, S. E., Parra, J. L., Jones, P. G. & Jarvis, A. Very high resolution interpolated climate surfaces for global land areas. International Journal of Climatology: A Journal of the Royal Meteorological Society 25, 1965–1978 (2005).
Mi, Q., Li, X. & Gao, J. How to improve the welfare of smallholders through agricultural production outsourcing: Evidence from cotton farmers in Xinjiang, Northwest China. Journal of Cleaner Production 256, 120636 (2020).
Gao, W. & Han, R. Xinjiang Statistical Yearbook - 2020. (China Statistics Press, 2020).
Li, W. & Yu, X. Xinjiang Production & Construction Corps Statistical Yearbook - 2020. (China Statistics Press, 2020).
Shafizadeh-Moghadam, H., Khazaei, M., Alavipanah, S. K. & Weng, Q. Google Earth Engine for large-scale land use and land cover mapping: an object-based classification approach using spectral, textural and topographical factors. GISci. Remote Sens. 58, 914–928, https://doi.org/10.1080/15481603.2021.1947623 (2021).
Haralick, R. M., Shanmugam, K. & Dinstein, I. H. Textural features for image classification. IEEE Transactions on systems, man, and cybernetics, 610–621 (1973).
Tkachev, G., Frey, S. & Ertl, T. S4: Self-supervised learning of spatiotemporal similarity. IEEE Transactions on Visualization and Computer Graphics (2021).
Altman, N. & Krzywinski, M. POINTS OF SIGNIFICANCE Ensemble methods: bagging and random forests. Nat. Methods 14, 933–934, https://doi.org/10.1038/nmeth.4438 (2017).
Wang, Q. et al. Assessment of calibration methods for nitrogen estimation in wet and dry soil samples with different wavelength ranges using near-infrared spectroscopy. Computers and Electronics in Agriculture 186, https://doi.org/10.1016/j.compag.2021.106181 (2021).
Kang, X. D., Li, S. T. & Benediktsson, J. A. Spectral-Spatial Hyperspectral Image Classification With Edge-Preserving Filtering. IEEE Transactions on Geoscience and Remote Sensing 52, 2666–2677, https://doi.org/10.1109/tgrs.2013.2264508 (2014).
Tomasi, C. & Manduchi, R. in Sixth international conference on computer vision (IEEE Cat. No. 98CH36271). 839–846 (IEEE).
Kang, X. et al. XJ_COTTON10: The 10-m cotton maps in Xinjiang, China during 2018-2021. Zenodo https://doi.org/10.5281/zenodo.7856467 (2022).
Ghorbanian, A. et al. Improved land cover map of Iran using Sentinel imagery within Google Earth Engine and a novel automatic workflow for land cover classification using migrated training samples. ISPRS Journal of Photogrammetry and Remote Sensing 167, 276–288, https://doi.org/10.1016/j.isprsjprs.2020.07.013 (2020).
Loosvelt, L. et al. Random Forests as a tool for estimating uncertainty at pixel-level in SAR image classification. International Journal of Applied Earth Observation and Geoinformation 19, 173–184 (2012).
Wang, S. & Wang, S. Land use/land cover change and their effects on landscape patterns in the Yanqi Basin, Xinjiang (China). Environmental monitoring and assessment 185, 9729–9742 (2013).
Li, X. et al. Land use/cover and landscape pattern changes in Manas River Basin based on remote sensing. International Journal of Agricultural and Biological Engineering 13, 141–152 (2020).
Wang, Y., Zang, L. & Araral, E. The impacts of land fragmentation on irrigation collective action: Empirical test of the social-ecological system framework in China. Journal of Rural Studies 78, 234–244 (2020).
Latruffe, L. & Piet, L. Does land fragmentation affect farm performance? A case study from Brittany, France. Agric. Syst. 129, 68–80 (2014).
Rahman, S. & Rahman, M. Impact of land fragmentation and resource ownership on productivity and efficiency: The case of rice producers in Bangladesh. Land use policy 26, 95–103 (2009).
Qi, L. et al. Climate and anthropogenic controls of seaweed expansions in the East China Sea and Yellow Sea. Geophysical Research Letters, e2022GL098185 (2022).
Zhu, X., Xiao, G., Zhang, D. & Guo, L. Mapping abandoned farmland in China using time series MODIS NDVI. Sci. Total Environ. 755, 142651 (2021).
She, B. et al. Identification and mapping of soybean and maize crops based on Sentinel-2 data. International Journal of Agricultural and Biological Engineering 13, 171–182, https://doi.org/10.25165/j.ijabe.20201306.6183 (2020).
Luo, C. et al. Using Time Series Sentinel-1 Images for Object-Oriented Crop Classification in Google Earth Engine. Remote Sensing 13, https://doi.org/10.3390/rs13040561 (2021).
Xilaer, R. & Yang, L. Monitoring Spatial-temporal Change of Cotton Phenology in Xinjiang and Its Response to Climate Change. Remote Sensing Technology and Application 33, 923–931 (2019).
Hughes, G. On the mean accuracy of statistical pattern recognizers. IEEE transactions on information theory 14, 55–63 (1968).
Hao, P., Tang, H., Chen, Z., Meng, Q. & Kang, Y. Early-season crop type mapping using 30-m reference time series. Journal of Integrative Agriculture 19, 1897–1911, https://doi.org/10.1016/s2095-3119(19)62812-1 (2020).
Wang, D., Qiu, P., Wan, B., Cao, Z. & Zhang, Q. Mapping α-and β-diversity of mangrove forests with multispectral and hyperspectral images. Remote Sensing of Environment 275, 113021 (2022).
Tan, J. et al. The flavonoid pathway regulates the petal colors of cotton flower. PLoS One 8, e72364, https://doi.org/10.1371/journal.pone.0072364 (2013).
Liang, Q. et al. A MYB transcription factor containing fragment introgressed from Gossypium bickii confers pink flower on Gossypium hirsutum L. Industrial Crops and Products 192, https://doi.org/10.1016/j.indcrop.2022.116121 (2023).
Jiang, Y. et al. DeepFlower: a deep learning-based approach to characterize flowering patterns of cotton plants in the field. Plant Methods 16, 156, https://doi.org/10.1186/s13007-020-00698-y (2020).
Wang, L. et al. Genetic mapping of fiber color genes on two brown cotton cultivars in Xinjiang. SpringerPlus 3, 1–5 (2014).
Sun, J., Sun, Y. & Zhu, Q. H. Breeding Next-Generation Naturally Colored Cotton. Trends Plant Sci 26, 539–542, https://doi.org/10.1016/j.tplants.2021.03.007 (2021).
Xu, W. et al. Cotton yield estimation model based on machine learning using time series UAV remote sensing data. International Journal of Applied Earth Observation and Geoinformation 104, https://doi.org/10.1016/j.jag.2021.102511 (2021).
Sun, S. et al. Three-dimensional photogrammetric mapping of cotton bolls in situ based on point cloud segmentation and clustering. ISPRS Journal of Photogrammetry and Remote Sensing 160, 195–207, https://doi.org/10.1016/j.isprsjprs.2019.12.011 (2020).
Siegfried, J. et al. Combining a cotton ‘Boll Area Index’with in-season unmanned aerial multispectral and thermal imagery for yield estimation. Field Crops Research 291, 108765 (2023).
Shi, G. et al. Cotton Yield Estimation Using the Remotely Sensed Cotton Boll Index from UAV Images. Drones 6, https://doi.org/10.3390/drones6090254 (2022).
Wang, L., Wang, J., Liu, Z., Zhu, J. & Qin, F. Evaluation of a deep-learning model for multispectral remote sensing of land use and crop classification. The Crop Journal (2022).
Chen, H. et al. Stacked spectral feature space patch: An advanced spectral representation for precise crop classification based on convolutional neural network. The Crop Journal https://doi.org/10.1016/j.cj.2021.12.011 (2022).
Xu, Y. et al. Artificial intelligence: A powerful paradigm for scientific research. The Innovation 2, 100179 (2021).
You, J., Li, X., Low, M., Lobell, D. & Ermon, S. in Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence (AAAI-17). 4559–4565 (2017).
Zang, H. et al. Short-term global horizontal irradiance forecasting based on a hybrid CNN-LSTM model with spatiotemporal correlations. Renewable Energy 160, 26–41 (2020).
Xu, J. et al. Towards interpreting multi-temporal deep learning models in crop mapping. Remote Sensing of Environment 264, https://doi.org/10.1016/j.rse.2021.112599 (2021).
Xu, J. et al. DeepCropMapping: A multi-temporal deep learning approach with improved spatial generalizability for dynamic corn and soybean mapping. Remote Sensing of Environment 247, https://doi.org/10.1016/j.rse.2020.111946 (2020).
Acknowledgements
This work was funded by Key Research Program of Frontier Sciences, CAS (grant number ZDBS-LY-DQC012), the National Natural Science Foundation of China (grant numbers 41971321 and 41830108), XPCC Science and Technology Project (grant number 2022CB002-01) and Open Fund of Key Laboratory of Oasis Eco-agriculture, XPCC (grant numbers 201801 and 202003). Changping Huang was supported by Youth Innovation Promotion Association, CAS (grant number Y2021047).
Author information
Authors and Affiliations
Contributions
Conceptualization: X.Y.K., C.P.H., L.F.Z.; Data curation: X.Y.K., T.Z., J.W.; Formal analysis: X.Y.K., C.P.H., J.M.C.; Funding acquisition: C.P.H., L.F.Z., Z.Z., X.L.; Investigation: X.Y.K., J.W., T.Z., H.H.W., X.L.F., Y.R.M., X.Y.; Methodology: X.Y.K., C.P.H., T.Z.; Software: X.Y.K., T.Z.; Supervision: C.P.H., L.F.Z., Z.Z., Q.X.T.; Validation: X.Y.K., T.Z., Z.Z.; Visualization: X.Y.K., T.Z.; Writing—original draft:.X.Y.K., C.P.H.; Writing—review & editing: J.M.C., L.F.Z., Q.X.T., Z.Z., X.L. Any questions on the dataset, please contact: Dr. Changping Huang, Ze Zhang, or Lifu Zhang.
Corresponding authors
Ethics declarations
Competing interests
The authors declare no competing interests.
Additional information
Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary information
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Kang, X., Huang, C., Chen, J.M. et al. The 10-m cotton maps in Xinjiang, China during 2018–2021. Sci Data 10, 688 (2023). https://doi.org/10.1038/s41597-023-02584-3
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41597-023-02584-3