Next Article in Journal
Smart Distribution Mechanisms—Part I: From the Perspectives of Planning
Previous Article in Journal
Regime Shifts in the Hexi Oases over the Past Three Decades: The Case of the Linze Oasis in the Middle Reaches of the Heihe River
Previous Article in Special Issue
Remote Sensing Land Use Evolution in Earthquake-Stricken Regions of Wenchuan County, China
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Deep Learning Semantic Segmentation Method for Landslide Scene Based on Transformer Architecture

1
Aerospace Information Research Institute, Chinese Academy of Sciences, Beijing 100094, China
2
Key Laboratory of Technology in Geo-Spatial Information Processing and Application System, Chinese Academy of Sciences, Beijing 100190, China
3
School of Electronic, Electrical and Communication Engineering, University of Chinese Academy of Sciences, Beijing 101408, China
4
State Key Laboratory of Remote Sensing Science, Aerospace Information Research Institute, Chinese Academy of Sciences, Beijing 100101, China
5
Institute of Artificial Intelligence, Beihang University, Beijing 100191, China
6
Satellite Application Center, Beijing 100094, China
7
Airlook Aviation Technology (Beijing) Co., Ltd., Beijing 100070, China
*
Authors to whom correspondence should be addressed.
Sustainability 2022, 14(23), 16311; https://doi.org/10.3390/su142316311
Submission received: 27 October 2022 / Revised: 1 December 2022 / Accepted: 2 December 2022 / Published: 6 December 2022

Abstract

:
Semantic segmentation technology based on deep learning has developed rapidly. It is widely used in remote sensing image recognition, but is rarely used in natural disaster scenes, especially in landslide disasters. After a landslide disaster occurs, it is necessary to quickly carry out rescue and ecological restoration work, using satellite data or aerial photography data to quickly analyze the landslide area. However, the precise location and area estimation of the landslide area is still a difficult problem. Therefore, we propose a deep learning semantic segmentation method based on Encoder-Decoder architecture for landslide recognition, called the Separable Channel Attention Network (SCANet). The SCANet consists of a Poolformer encoder and a Separable Channel Attention Feature Pyramid Network (SCA-FPN) decoder. Firstly, the Poolformer can extract global semantic information at different levels with the help of transformer architecture, and it greatly reduces computational complexity of the network by using pooling operations instead of a self-attention mechanism. Secondly, the SCA-FPN we designed can fuse multi-scale semantic information and complete pixel-level prediction of remote sensing images. Without bells and whistles, our proposed SCANet outperformed the mainstream semantic segmentation networks with fewer model parameters on our self-built landslide dataset. The mIoU scores of SCANet are 1.95% higher than ResNet50-Unet, especially.

1. Introduction

Landslide [1] is a geological phenomenon with great danger. The occurrence of landslide is caused by both natural and human factors. Natural factors mainly include terrain, lithology, geological structure, bad weather, etc. And human factors are mainly human activities that violate the laws of nature and destroy the stable conditions of slopes. Landslides cause great damage to industrial and agricultural production as well as people’s lives and properties. In severe cases, landslides even cause devastating disasters. For instance, in October 2021, landslides in Northeast and Southwest India caused massive casualties, infrastructure damage, crop damage and other serious disasters. After a landslide occurs, so as to facilitate rescue operations and ecological restoration work, it is very important to use satellite data or aerial photography data to quickly locate and estimate the area of the landslide [2]. In recent years, with the rapid development of remote sensing technology, more and more high-resolution remote sensing images [3] can be obtained. With rich information and high resolution, remote sensing images are gradually playing a more important role in various fields of national life. For example, in landslide disasters, remote sensing images are used to assess the area and extent of landslide impact. In order to identify landslide hazards and perform further analysis and processing, we need specific methods to separate and extract regions of interest from remote sensing images. At the same time, various remote sensing image vision tasks based on deep learning [4] methods have been greatly promoted and developed. Specifically, remote sensing image segmentation [5] can complete the pixel-level prediction of the image to effectively obtain image information. For landslide scenes, deep learning semantic segmentation methods accurately identify landslide areas to carry out disaster relief work. It is very suitable and great to solve the above regional positioning and area estimation.
Deep learning is used to learn the underlying distribution and representation level of sample data. The goal of deep learning is to make machines have the ability to analyze problems and learn knowledge like humans. Deep learning, a data-driven machine learning algorithm, has made outstanding progress in many fields, such as video scene [6] and vision scene [7]. However, deep learning methods have not been applied deeply enough in the field of various natural disasters. So as to achieve the effect of controlling, managing and reducing disasters, using deep learning technology [8] to predict and evaluate landslide areas can quickly and accurately obtain space disaster information. Taking landslide disasters in the Loess Plateau as the research object, to complete the regional positioning and area estimation of landslides, we used semantic segmentation technology based on deep learning to process and analyze remote sensing landslide images.
There are various segmentation methods based on deep learning. The mainstream architecture of semantic segmentation methods is the Encoder-Decoder [9] architecture. In order to obtain the high-level and low-level semantic information, we use the encoder to extract the features of the original image. At present, universal encoder networks are the Convolutional Neural Network [10] (CNN) based on convolution operation and the Transformer [11] self-attention mechanism [12]. Many experiments [13] showed that Transformer has a stronger ability than the convolutional neural network to extract image features. The strong performance of Transformer is attributed to the self-attention mechanism capturing global information. However, due to the self-attention mechanism causing the existence of high computational complexity, Transfromer is not widely used. The decoder network is used to fuse high-level and low-level semantic information [14] obtained by the encoder. In the decoder stage, through processing the low-level feature information after downsampling in the encoder stage to extract rich high-level feature semantic information, then, through related techniques, the corresponding feature information is adjusted to the resolution of the input image to complete pixel-level prediction. Currently, there are still very few pixel-level landslide labeled datasets. Due to the fact that pixel-by-pixel labeling of landslide datasets requires a lot of labor and financial costs, conducting experiments and testing effect in landslide scenarios by using deep learning semantic segmentation methods is difficult.
Faced with the above problems, the main contributions of this paper are as follows:
  • We construct a landslide dataset based on landslide remote sensing Image in the Loess Plateau. We use support vector machines to annotate the remote sensing images of landslides to get preliminary label data. By image post-processing and manual correction, we obtained a well-labeled landslide dataset.
  • On the existing landslide dataset, we conduct related experiments on different and representative semantic segmentation network. After that, we compare and analyze the performance of different networks.
  • We propose a deep learning semantic segmentation method based on Encoder-Decoder architecture for landslide recognition, called Separable Channel Attention Network (SCANet). SCANet consists of two parts, Poolformer as the encoder and Separable Channel Attention Feature Pyramid Network (SCA-FPN) as the decoder. Poolformer is based on transformer improvement. SCA-FPN is our uniquely designed feature pyramid network. Final experiments show that our method is better than the existing representative semantic segmentation networks on the landslide dataset.

2. Materials

2.1. Dataset Source

The dataset of images of landslides in this paper are derived from high-resolution remote sensing images and landslide datasets based on terrain interpretation. The contents of the images mainly cover the landslide area of the Loess Plateau.

2.2. Dataset Annotation

The landslide dataset used in this paper only contains 500 remote sensing images. There are no pixel-level annotations for the landslide areas in these images. Our annotation process, shown in Figure 1, can be divided into three steps: Pre-labeling, Post-processing, Manual correction.

2.2.1. Pre-Labeling

We used support vector machine [15] (SVM) to complete the pre-annotation of the image. SVM is a machine learning algorithm that uses supervised learning methods for the binary classification of data. The learning strategy of SVM is to maximize the interval, which can transform the algorithm process into solving the convex optimal quadratic programming problem. It does not need to rely on overall data, but solves small sample machine learning problems well.
Our specific approach was as follows. On the landslide dataset, for each image, we first manually selected some small rectangular areas that can represent the landslide in the image, then selected some small rectangular areas that represented the background in the picture. It means that, except for the landslide area, other areas are regarded as the background. These areas are support vectors to update the model parameters of the optimized support vector machine to complete the pre-labeling of the landslide dataset.

2.2.2. Post-Processing

We used principal or second components analysis [16] to post-process the labeled images. Principal components analysis (PCA) is a method of post-processing similar to convolution filtering. It classifies all pixels in the region according to which area corresponds to the transform kernel size into principal category. Similarly, second components analysis (SCA) classifies all pixels in the region according to which area corresponds to the transform kernel size into principal category. The formula is as follows:
C x , y = F ( x , y ) = C p r i , ( x , y ) R , P C A , C s e c , ( x , y ) R , S C A .
where R represents the image area of the same size as the transform kernel, ( x , y ) represents the coordinate position of the pixel, F represents the transform kernel algorithm, C p r i represents principal components, C s e c represents second components, C x , y represents the category in which the image area is classified.

2.2.3. Manual Correction

Labels obtained through support vector machines and image post-processing methods had several label errors. For this part, we manually corrected this part to obtain the final labeled landslide dataset. With the use of machine learning methods for pre-labeling, manual correction only requires very little human and financial resources compared to manual labeling.

2.3. Dataset Preprocessing

The image sizes in our landslide dataset were different. There were too many pixels in a single remote sensing landslide image. In order to facilitate model training, we cropped the remote sensing image to make it reach a fixed size. At the same time, deep neural networks often require large amounts of data to train to avoid overfitting. To solve this problem, we used data augmentation to increase the diversity of the landslide dataset to better train neural network models.

2.3.1. Image Cropping

We crop the image to a fixed size. For the training set, we use the smooth cropping method to crop the remote sensing image and the corresponding annotation data to a size of 256 × 256 . For the problem of boundary continuity, the overlap ratio was set to 0.25 during the cropping process. In order to ensure that the number of foreground pixels and background pixels does not differ much, we remove the landslide image where the ratio of landslide pixels is too large or small and only keep the landslide image with the ratio of landslide pixels in the range of 0.05–0.9. For the test set, since the test process does not modify the model weights, the image cropping operation can be omitted.

2.3.2. Data Augmentation

The data augmentation method [17] for image semantic segmentation is similar to other computer vision tasks. The methods we used are mainly as follows:
  • Flip transformation;
  • Color dithering;
  • Contrast transformation;
  • Noise perturbation;
  • Rotation transformation.
For color dithering, contrast transformation and noise perturbation, the label corresponding to the remote sensing image does not change. For flip transformation and rotation transformation, the label changes as the image changes. It can be seen in Figure 2 that the raw image has some changes in color dithering and rotation transformation.

3. Related Works

Semantic segmentation [18] is a classic visual scene problem, where the vision task is to take raw image data as input and transform them into masks with salient interests. According to the object, each pixel in the raw image data is assigned to a specified category it belongs to. The semantic segmentation task can provide pixel-level image understanding in a completely human-perceived way. It combines visual tasks such as image classification and object detection. Semantic segmentation divides the image into regional blocks with certain semantic meaning by a specific method and identifies the semantic category of each regional block. It implements the process of the inference from low-level semantics to high-level semantics and finally obtains segmented images with pixel-by-pixel annotations. At present, image semantic segmentation methods include traditional machine learning methods [19] and modern deep learning methods [20]. Traditional semantic segmentation methods can be divided into statistical-based methods [21] and geometric-based methods [22]. With the continuous development of artificial intelligence, the semantic segmentation method of deep learning greatly surpasses the traditional semantic segmentation method. Compared with traditional semantic segmentation methods, deep learning methods use neural networks to automatically learn image features and directly complete end-to-end learning tasks. A large quantity of image semantic segmentation experiments showed that deep learning methods perform better in improving the accuracy of semantic segmentation. The current mainstream end-to-end semantic segmentation networks based on deep learning are encoder-decoder structures, which is shown in Figure 3. The encoder extracts the features of the original image, and the decoder completes the fusion of information based on these features, thereby completing the pixel-by-pixel prediction of the original image.

3.1. Mainstream Encoder Networks

3.1.1. Convolutional Neural Network

Convolutional neural network [23] is a kind of feedforward neural network [24] including convolutional computation, and has the ability of representation learning. While ensuring translation invariance, convolutional neural network can process input data according to its hierarchical structure. The structural characteristics of convolutional neural networks are local area connection, weight sharing, and downsampling. These characteristics effectively reduce the number of parameters of the network and alleviate the overfitting problem of the model. The main structure of the convolutional neural network is as follows:
  • Convolutional layer
    Each convolutional layer consists of several convolution kernels. The parameters of each convolution are obtained through the back-propagation algorithm. The purpose of the convolution operation is to extract different features of the input data. Shallow convolution can extract low-level features such as edges, lines, and corners. Deep convolution can extract more complex high-level features.
  • Rectified Linear Units layer
    This layer needs to use an activation function [25]. The activation function activates a certain part of the neurons in the neural network and transmits the activation information to the next layer of the neural network. Activation functions are generally non-linear functions. The reason why neural networks can solve non-linear problems is that they introduce a non-linear activation function which makes up for the expressive ability of the linear model.
  • Pooling layer
    After the convolutional layer, features with larger dimensions will be obtained. The pooling operation can divide the features into several regions. Then, it performs some operations, such as taking maximum value or average value, to obtain new, smaller dimensional features. The pooling operation [26] can achieve a nonlinear effect and expand the receptive field. The pooling operation also has the invariance of translation, rotation and scale.
  • Fully-Connected layer
    The function of this layer is to integrate the semantic information output by each block, which combines local information into global information to calculate the final classification score. When the convolutional neural network is used as an encoder, the fully connected layer [27] will be removed.
At present, there are various variants of convolutional neural networks. Representative networks include VGGNet [28], ResNet [29], ConvNeXt [30], etc. Lightweight convolutional networks [31] include ShuffleNet [32], MobileNet [33], EfficientNet [34], etc.

3.1.2. Transformer Architecture

Before the advent of Transformer [11], the mainstream networks in natural language processing [35] were based on recurrent or convolutional neural networks. The recurrent neural network [36] connected by the attention mechanism has the best performance. Recurrent neural network is a sequential model [37] which cannot solve the problem of long dependencies [38]. When the sequence of input data is too long, in the process of data processing by the sequence model, the information will be gradually lost. At the same time, it is also difficult to perform parallel computing in the sequential model. Transformer is a simple model that abandons the neural network structure of recurrent and convolution and only relies on the attention mechanism. Transformer introduces a self-attention mechanism that makes the modeling of dependencies independent of the input and output sequences, which solves the problem of long-distance dependencies and supports parallel computing.
By the end of 2020, Transformer had shown a revolutionary improvement in the field of computer vision. Transformer architecture surpassed the performance of the convolutional neural network, which often topped the vision list in many fields. This also showed that computer vision and natural language processing were expected to be unified under the Transformer [13] architecture. The power of the Transformer network relies on the self-attention mechanism. Its main structure mainly includes:
  • Self-attention
    The attention mechanism formula is as follows:
    A t t e n t i o n ( Q , K , V ) = S o f t m a x ( Q K T d k )
    where the matrix Q, K, V has the same dimensions N × C , N = H × W , which represent the length of the sequence, C represents the embedding dimensions, where 1 d k is the scaling factor.
    Q , K , V = L i n e a r ( X ) , L i n e a r ( X ) , L i n e a r ( X )
    where X represents the input features. Self-attention does linear mapping based on the attention mechanism to obtain the matrix Q , K , V .
    M u l t i H e a d ( Q , K , V ) = L i n e a r ( C o n c a t ( h e a d 1 , h e a d 2 , . . . , h e a d n ) ) w h e r e h e a d i = A t t e n t i o n ( Q i , K i , V i )
    where Q , K , V is split into Multi-Head Q 1 , 2 , . . . , n , K 1 , 2 , . . . , n , V 1 , 2 , . . . , n . By calculating the attention of multiple heads, it will be found that each of the different channels in the space of attention is different. While the computational complexity of the model is similar, the representation ability of the model is improved.
  • Positional Encodings
    Since Transformer contains no recurrence and no convolution, in order for Transformer to make use of the order of the sequence, Transformer is injected with some information about the relative or absolute position of the tokens in the sequence. The positional encodings have the same dimension as the embeddings. There are many choices of positional encodings [39]. However, Transformer uses sine and cosine functions of different frequencies:
    P E p o s , 2 i = s i n ( p o s / 10000 2 i / d m o d e l ) P E p o s , 2 i + 1 = c o s ( p o s / 10000 2 i / d m o d e l )
    where pos is the position, i is the dimension and d m o d e l is the embedding dimensions. That is, each dimension of the positional encoding corresponds to a sinusoid.
At present, there are applications based on transformers in three major visual fields of classification, detection and segmentation. Representative networks include ViT [40], Mix Transformer [41], Swin Transformer [42], etc.

3.2. Mainstream Decoder Networks

3.2.1. Symmetrical Architecture

Symmetric network [43] can actually be regarded as a codec structure. Representative networks include UNet [44] and LinkNet [45], etc. UNet is a U-shaped symmetric structure with convolutional layers on the left and upsampling layers on the right. When implementing, we can design the network from scratch and initialize the weights. After that, we train the model, also can use the existing network and load the corresponding trained weight files, then we build the upsampling layers for training calculation. Lineknet draws on the idea of U-Net, and the innovation of it lies in the connection between the encoder and the decoder. After multiple downsampling by the encoder, spatial information is lost partly. It is difficult to restore the lost spatial information in the decoder part, so the input and output of the encoder are sent together to the network for training.

3.2.2. Multi-Scale Analysis

Multi-scale analysis is a representative method in image processing which has been widely used in various neural networks. The specific method is to use the inherent multi-scale pyramid hierarchy of deep convolutional neural networks to construct feature pyramids with marginal additional cost. Currently, there are many variants of feature pyramid networks [46], such as the Pyramid Scene Parsing Network [47] (PSPNet). It is a multi-scale network that can better learn global contextual representations of scenes. PSPNet uses a residual network as a feature extractor to extract different feature maps. Then, according to different size patterns, these features are mapped into the pyramid module. Each scale-sized feature map corresponds to a pyramid layer. At the same time, these feature maps are processed to reduce the dimensions by a 1 × 1 convolutional layer. The output of the pyramid is upsampled and concatenated with the initial feature maps to capture local and global contextual information. Finally, pixel-wise prediction is finished by using a softmax layer.

3.2.3. DeepLab Based on Dilated Convolution

Dilated convolution [48] introduces dilated rate in the convolutional layer. It can enlarge the receptive field without increasing the computational cost. The Deeplabv2 [49] network uses dilated convolution to solve the problem of reduced resolution in the network caused by max pooling and striding. The key structure of Deeplabv2 network is Atrous Spatial Pyramid Pooling (ASPP). To classify the center pixel, ASPP exploits multi-scale features by employing multiple parallel filters with different rates. Deeplab-ASPP captures object and image context at multiple scales to reliably segment objects at multiple scales. It combines the methods of deep CNN networks and probabilistic graphical models to improve the localization of object boundaries. On this basis, Deeplabv3 [50] proposes a more general framework which is suitable for semantic segmentation tasks in more scenarios. The Deeplabv3 model can control feature extraction and learn network structure of multi-scale features. In the Deeplabv3 model, based on the pre-trained ResNet, the last ResNet block uses dilated convolution. At the same time, it uses different hole rates to obtain multi-scale information, and the decoding part also uses Atrous Space Pyramid Pooling.

4. Methods

The framework of our Separable Channel Attention Network (SCANet) is based on mainstream Encoder-Decoder architectures, which are illustrated in Figure 4. The framework consists of two parts, Poolformer [51] as the encoder and Separable Channel Attention Feature Pyramid Network (SCA-FPN) as the decoder. Firstly, Poolformer is improved on the basis of Transfromer architecture. Poolformer replaces the self-attention mechanism in Transformer with a pooling operation. This replacement greatly reduces the complexity of the network while maintaining very good performance. Secondly, SCA-FPN is a feature pyramid structure. We inserted a separable channel attention module that we designed originally into SCA-FPN. Separable channel attention includes spatial attention and channel attention. SCA-FPN can fuse different levels of spatial information and channel information obtained by separable channel attention. At the same time, Separable channel attention is an independent module in the calculation process, so that it can be easily embedded and used in other networks. The overall SCANet network, we designed, is spliced by Poolformer and SCA-FPN, and it exhibits better performance with reduced network computational complexity.

4.1. Poolformer Encoder

Poolformer adopts the same general framework as Transformer. Its structure is shown in Figure 4. Poolformer has the ability to extract multi-scale information. Given an input image I R H × W × C , we fed it into Poolformer to extract features to obtain multi-level feature maps C i ( i = 1 , 2 , 3 , 4 ) in size S i 1 4 , 1 8 , 1 16 , 1 32 of the original image resolution. Poolformer has four important components, including Patch Embedding, Layer Normalization, Residual Connection and Toekn Mixers.

4.1.1. Patch Embedding

The function of Patch Embedding [52] is encoding the input image to adapt the input interface of Poolformer. It cuts an input image into a series of image blocks which have the same size, then encodes the image blocks by convolution to obtain image embedding. Convolution kernel is the same as the size of the image block. In order to ensure the continuity between the image blocks, we use an overlap cutting method in the process of converting the image into image blocks. Another function of Patch Embedding is to downsampling the feature map between Poolformer blocks. It means that each Poolformer block has Patching Embedding. Regardless of continuity, Patch Embedding can be expressed as:
Y H o × W o = C o n v K × K ( X H i × W i ) Z N = I n p u t E m b e d ( Y H o × W o )
where Y represents the output feature, X represents the input feature, H o = H i k , W o = W i k , N = H o × W o .

4.1.2. Layer Normalization

In the process of using the gradient descent algorithm to optimize model parameters, with the deepening of the network depth, data distribution will have changed. In order to ensure the stability of data distribution and prevent the occurrence of gradient explosion, it is necessary to normalize the data transmitted in the network. Batch Normalization [53] is usually used in convolutional neural networks to normalize data in the dimension of batches. It can balance the data distribution and speed up the convergence of the network. However, for modeling problems of uncertain length sequences, Batch Normalization cannot be embedded in the network to use. Poolformer is actually an indeterminate long sequence modeling network, so it adds Layer Normalization [54] to each Poolformer block rather than Batch Normalization. Layer Normalization prevents gradient diffusion and speeds up parameter convergence for Poolformer. Different from Batch Normalization, Layer Normalization calculates the mean and variance in the channel dimension to normalize the data. The specific formula is as follows:
μ l = 1 H i = 1 H x i l σ l = 1 H i = 1 H ( x i l μ l ) 2 y l = x l μ l σ l + ϵ γ + β
where l represents the number of layers of the neural network, H represents the number of hidden units in the layer, ϵ is the bias that prevents standard deviation from being zero, γ and β are linear affine transformation parameters.

4.1.3. Residual Connection

Poolformer is a network with deep layers like Transformer. As the number of neural network layers increases, semantic information at different levels in Poolformer can be extracted. After obtaining a large amount of shallow and deep semantic information, we will have more ways to fuse this semantic information to make more accurate predictions. However, too-deep neural network layers will lead to some problems, such as vanishing gradients and exploding gradients. With the number of network layers increasing, the characteristics of the neural network also change unpredictably. The performance of a deep network may be worse than that of a shallow network. Residual structure can solve the problem of network degradation, vanishing gradients and exploding gradients very well. In Poolformer, adjacent layers are connected through a residual structure. Residual connection is defined as the superposition of the input and the nonlinear change of the input. The formula of the residual connection is as follows:
x l + 1 = f ( h ( x l ) + F ( x l , W l ) )
where l represents the position of the network layer, W represents the weight of the network layer, h , F , f are short-cut mapping, residual mapping, activation mapping.

4.1.4. Token Mixers

MetaFormer [51]’s components are similar to Transformer except for token mixer. MetaFormer is a general architecture where the token mixer is not specified, which is illustrated in Figure 5. For example, TokenMixer is replaced with a pooling operation in Poolformer. Embedding tokens that come from Patch embedding X are fed to Metaformer blocks. Each Meataformer block consists of two residual sub-blocks. The first sub-blocks use the token mixer to communicate information from embedding tokens. It can be expressed as:
Y = T o k e n M i x e r ( L N ( X ) ) + X
where L N ( · ) represents Layer Normalization. T o k e n M i x e r ( · ) represents a module that can work for mixing token information such as the self-attention mechanism in vision Transformer models, spatial MLP in MLP-like models [55] and the pooling operation in Poolformer.
The second sub-blocks use two-layer MLP with non-linear activation to communicate information from token mixer. It can be expressed as
Z = σ ( L N ( Y ) W 1 ) W 2 + Y
where W 1 R C × C h i d d e n and W 2 R C h i d d e n × C are linear affine transformation parameters. σ ( · ) represents a non-linear activation, such as ReLU, GELU, SiLU.
Compared with Transformer, Poolformer removes Transformer’s self-attention mechanism. The main difference made by Poolformer is using simple pooling as a token mixer. For input data T R ( C × W × H ) , the pooling operation is expressed as
T i , j = 1 K × K p , q = 1 K T i + p K + 1 2 , j + q K + 1 2 T i , j
where K is the pooling size.

4.2. SCA-FPN

SCA-FPN is the decoder of SCANet we designed. The function of SCA-FPN is to fuse semantic features of different levels obtained by Poolformer encoder to complete the pixel-level prediction of the original image. SCA-FPN has two important components, including Separable Channel Attention and Feature Pyramid Network.

4.2.1. Separable Channel Attention

The thought of Separable Channel Attention (SCA) is to focus on different information in different dimensions. Separable channel attention module, shown in Figure 6, divides semantic features into spatial dimensions and channel dimensions. Half of the semantic features are used to focus on spatial information, and half of the semantic features are used to focus on channel information. We use the full convolution operation to get spatial information and use convolution and pooling operations to get channel information. The final feature map is obtained by splicing spatial information and channel information. The specific implementation formula is as follows:
Y 1 s = C o n v C 2 × W × H ( X ) , Y 1 c = C o n v C 2 × W × H ( X ) Y 2 s = C o n v 1 × W × H ( Y 1 s ) , Y 2 c = P o o l W × H ( Y 1 c ) Z = C o n c a t ( Y 1 s × Y 2 s , Y 1 c × Y 2 c ) + X
where s , c represent spatial information and channel information, respectively.

4.2.2. Feature Pyramid Network

Feature Pyramid Network [46] (FPN) is a structure based on multi-scale analysis. The overall structure of SCA-FPN is a feature pyramid network, which is shown in Figure 4, to fuse low-resolution and high-resolution features. Feature Pyramid Network consists of bottom-up paths, top-down paths and lateral connections.
The bottom-up process is a normal forward propagation process of the neural network. The feature map usually becomes smaller and smaller after being calculated by the convolution kernel. The top-down process is used to upsample more abstract and semantically stronger high-level feature maps. Lateral connection is to merge feature maps obtained in the process of bottom-up and top-dowm. Firstly, we double upsample the low-resolution feature map, and the sampling method is nearest neighbor upsampling. Secondly, we merge the upsampled map with the corresponding bottom-up map by element-wise addition. The overall process is an iterative algorithm.
In our designed SCA-FPN decoder, The fusion method of feature maps is no longer a simple lateral connection. We inserted a SCA module in the laterally connected part of the network to produce the output of each stage. For semantic segmentation task, it uses a two-layer multilayer perceptron at the end of the network to generate masks. Finally, it generates predicted results by the way of upsampling.

4.3. Loss Function

Loss function is used to measure the degree of inconsistency between the predicted value of the model and the real value. In the training phase, our SCANet uses the standard cross-entropy loss and Dice loss [56] as the loss function. For the final predicted output F and the ground truth G, the formula is as follows:
Cross-Entropy Loss
l o s s ( F , G ) = 1 N k = 1 N [ G k l o g ( F K ) + ( 1 G k ) l o g ( 1 F k ) ]
Dice Loss
l o s s ( F , G ) = 1 2 | F G | | F | + | G |
Total Loss
l o s s = 0.5 l o s s c e + 0.5 l o s s d i c e
where k is the index of pixels and N is the number of pixels in F.

5. Experiments and Discussion

In this section, we conduct extensive experiments on the landslide dataset, which was mentioned in Section 2 to evaluate the the performance of our proposed SCANet. The details of the experimental setup are in Section 5.1. The comparison experiments and analysis of SCANet and mainstream semantic segmentation networks on the landslide dataset are provided in Section 5.2. The ablation experimental results and analysis on Poolformer and separable channel attention module are presented in Section 5.3. Overall effectiveness analysis of SCANet and mainstream semantic segmentation networks is provided in Section 5.4.

5.1. Experimental Settings

5.1.1. Implementation Details

We divide the landslide dataset into a training set and a test set. Due to the small number of data sets, the test set and validation set are the same. In the process of training the model, we perform data augmentation operations on the training set. The specific implementation is is shown in Table 1.
The SCANet is implemented in the PyTorch framework, trained and tested on a platform with a single NVIDIA GeForce RTX 3060(12 GB RAM) with CUDA version 10.3 and Cudnn version 8.2.0. On the landslide dataset, we randomly cropped 256 × 256 patches from the original image and randomly mirrored and rotated them with specified angles ( 0 , 90 , 180 , 270 ). The stochastic gradient descent with momentum (SGDM) optimizer with a momentum of 0.9 and an initial learning rate of 0.001 was set to guide the optimization. During the network training, a poly learning rate policy was adopted to adjust the learning rate. The batch size was set to 16 and the total number of training epochs was set to 300. Experimental settings are shown in Table 2.

5.1.2. Comparison Methods and Evaluation Metrics

Intending to fully prove the performance of the proposed SCANet for semantic segmentation of landslide remote sensing images, we compare the proposed SCANet with ten mainstream semantic segmentation networks based on Encoder-Decoder architecture, including MobileNetv2-DeepLabv3Plus [50,57], MobileNetv2-UNet [44,57], MobileNetv2-FPN [46,57], MobileNetv2-PSPNet [47,57], MobileNetv2-LinkNet [45,57], ResNet50-DeepLabv3Plus [29,50], ResNet50-UNet [29,44], ResNet50-FPN [29,46], ResNet50-PSPNet [29,47] and ResNet50-LinkNet [29,45].
In order to fairly compare our proposed SCANet with mainstream semantic segmentation methods on the landslide dataset, we use the widely used evaluation metrics as follows:
IoU
I o U i = x i , i j = 1 n x i , j + j = 1 n x j , i x i , i m I o U = 1 n i = 1 n I o U i
where x i , j means the number of instances of class i predicted as class j, and n is the number of classes.
Accuracy
O A = T P + T N P + N
where O A is the ratio of the number of correctly predicted pixels to the total number of pixels.
F1-score
F 1 = 2 × p r e c i s i o n × r e c a l l p r e c i s i o n + r e c a l l
where p r e c i s i o n = T P T P + F P and r e c a l l = T P T P + F N .

5.2. Comparative Experiments

We follow the experimental setup in Section 5.1 and conduct extensive experiments on the landslide dataset to compare the performance of our proposed SCANet and mainstream semantic segmentation networks, including MobileNetv2-DeepLabv3Plus [50,57], MobileNetv2-UNet [44,57], MobileNetv2-FPN [46,57], MobileNetv2-PSPNet [47,57], MobileNetv2-LinkNet [45,57], ResNet50-DeepLabv3Plus [29,50], ResNet50-UNet [29,44], ResNet50-FPN [29,46], ResNet50-PSPNet [29,47] and ResNet50-LinkNet [29,45].
The quantitative results of our comparative experiments on the landslide dataset are shown in Table 3. The visual analysis of evaluation metrics is shown in Figure 7, the visualization results of our proposed SCANet and mainstream semantic segmentation networks using Mobilenet_v2 as encoder are shown in Figure 8 and the visualization results of our proposed SCANet and mainstream semantic segmentation networks using ResNet50 as encoder are shown in Figure 9. Our proposed SCANet achieve SOTA in all semantic segmentation methods mentioned above.
As can be observed from Table 3 and Figure 7, our method achieves the best results on the evaluation metrics of Precision, OA, F1-score and IoU. Compared with mainstream semantic segmentation networks that use Mobilenet_v2 [57] as an encoder, though the amount of our model parameters increased, our method performed well. Specifically, our proposed SCANet outperformed the second-best method, Mobilenet_v2-Unet [44,57], by 3.29% and the third-best method, Mobilenet_v2-DeepLabV3Plus [50,57], by 4.39% in the IoU score. Compared with mainstream semantic segmentation networks that use ResNet50 [29] as an encoder, our method performed well while the amount of our model parameters decreased. Specifically, our proposed SCANet outperformed the second-best method, ResNet50-Unet [29,44], by 1.95% and the third-best method, ResNet50-DeepLabV3Plus [29,50], by 3.25% in the IoU score.
In addition, we conduct many detailed visual comparison experiments, further confirming the performance of the proposed SCANet for semantic segmentation tasks on the landslide dataset. The visualization results of the proposed SCANet and mainstream semantic segmentation networks that use Mobilenet_v2 [57] as an encoder are shown in Figure 8. The visualization results of the proposed SCANet and mainstream semantic segmentation networks that use ResNet50 [29] as an encoder are shown in Figure 9. Benefitting from Poolformer encoder, which effectively transfers global information to each pyramid-level feature map, SCANet can generate high-resolution feature maps with high-level semantic information. Benefitting from SCA-FPN decoder, which introduced separable channel attention, SCANet can predict the edge texture information of the image more accurately. The combination of Poolformer encoder and SCA-FPN decoder makes our method achieve the best performance.

5.3. Ablation Experiments

In this subsection, we evaluate the effectiveness of two key modules of our proposed SCANet based on Encoder-Decoder architecture, Poolformer as the encoder module, Separable Channel Attention used in SCA-FPN. The ablation experiments are also trained and tested on the landslide dataset. To verify the effect of Poolformer encoder, we conduct extensive ablation experiments in Section 5.3.1 that compare Poolformer with ResNet50 while keeping the decoder consistent. To verify the effect of Separable Channel Attention, we conduct extensive ablation experiments in Section 5.3.2 that compared SCA-FPN with FPN while keeping the encoder consistent.

5.3.1. Effect of Poolformer

Poolformer encoder can be spliced with many decoding modules flexibly, such as Unet [44], FPN [46], PSPNet [47], LinkNet [45]. Poolformer encoder was used to obtain pyramid-level features, and the final output was expressed as:
O u t = U p ( U p ( U p ( U p ( C 1 ) + C 2 ) + C 3 ) + C 4 )
where C i ( i = 1 , 2 , 3 , 4 ) represents pyramid-level features, U p represents upsampling.
As can be seen in Table 4, semantic segmentation networks using Poolformer encoder had better effects than networks using ResNet50 encoder, improving the IoU by 2.79%, 1.97%, 0.73%, 1.24% 2.91%, the F1-score by 1.66%, 1.21%, 0.82%, 0.76% 2.02% while using Unet [44], FPN [46], PSPNet [47], LinkNet [45], SCA-FPN as decoders, respectively. The performance improvement is due to Poolformer encoder, which can capture global information well. Compared with other decoders, the method using SCA-FPN decoder has the largest performance increase, improving the IoU by 2.79%, the F1-score by 1.66% after replacing ResNet50 with Poolformer. It indicates that the SCA-FPN decoder that we designed is more suitable for Poolformer encoder. All in all, Poolformer encoder actually provides a significant performance improvement for landslide scene segmentation.

5.3.2. Effect of SCA

We evaluate the effectiveness of our proposed SCA module by comparing FPN with SCA-FPN while keeping the encoder consistent. As is shown Table 5, the landslide segmentation performance is improved after adding SCA. ResNet50-SCA-FPN improves the IoU score by 1.28% compared to ResNet50-FPN while both networks use ResNet50 encoder. SCANet improves the IoU score by 2.22% compared to Poolformer-FPN while both networks use Poolformer encoder. Figure 10 shows four models’ heatmap that derived from the features before the network classification layer. We can see the network’s attention to the landslide area is enhanced after replacing FPN with SCA-FPN. It indicates the insertion of SCA makes the model pay more attention to the edge region and texture of the landslide area. Therefore, SCA is surely an effective module for semantic segmentation network in the landslide scene.

5.4. Analysis of Methods

We test the current mainstream semantic segmentation methods in the landslide scene. On our landslide dataset, experiments are carried out to evaluate the performance of each method. Comparing different SOTA networks, the neworks (MobileNetv2-UNet [44,57], ResNet50-UNet [29,44]) using Unet decoder performs the best. Unet decoder, as a representative of a lightweight decoder, achieves the best performance with combinations of different encoders.
Our proposed SCANet uses Poolformer as an encoder and SCA-FPN as a decoder. Unlike convolutional neural networks, Poolformer is based on Transformer architecture. In Poolformer, self-attention mechanisms in the network are replaced by pooling layers, which ensure that the model computational complexity is low. Compared with ResNet50 encoder, Poolformer encoder performs better while the model complexity is reduced. Besides, based on the FPN decoder, we embed our own designed SCA module to build a new SCA-FPN decoder for feature fusion and pixel prediction. From the ablation experiments, we can see that the SCA-FPN decoder we designed is better than the FPN decoder. SCA-FPN introduces a separable channel attention module to make the landslide area more focused. In short, compared with the mainstream semantic segmentation network mentioned above, our proposed SCANet performs best in the task of semantic segmentation of landslide scenes.

6. Conclusions

In this paper, based on the remote sensing images of landslides on the Loess Plateau, we use machine learning methods to construct a dataset of landslide scenes. In order to compare the performance of current mainstream semantic segmentation networks, we do relevant experiments on the landslide dataset for analysis. Unlike convolutional neural networks, we propose a new framework for semantic segmentation of remote sensing images named Separable Channel Attention Network (SCANet), which relies on the transformer architecture. SCANet contains two components, the Poolformer encoder and the SCA-FPN decoder. In the encoder part, for the convolutional neural network, the trained network cannot capture the global mutual information. However, Poolformer, which we use, makes up for the lack of the convolutional neural network. Limited by the high complexity of the self-attention algorithm in Transformer, Poolformer replaces self-attention mechanisms with pooling operations. This change still maintains network performance better than the convolutional neural network. In the decoder part, in order to make the network pay more attention to the image edge texture information, SCA-FPN uses a feature pyramid structure to obtain multi-scale information. The separable channel attention mechanism we designed is also inserted into SCA-FPN, which makes the network pay more attention to the foreground information and improves the accuracy of pixel-level classification.
In addition, we conduct extensive experiments on the landslide dataset. Through these experiments, we demonstrate that our SCANet can achieve good segmentation results on the semantic segmentation task of the remote sensing images. Our network outperforms other mainstream methods on the landslide dataset. Our research also validates that semantic segmentation techniques can be used to locate and estimate landslide areas. We hope this research can inspire more researchers in this area and deploy practical applications.

Author Contributions

Y.L. provided the landslide data; Z.W. and T.S. proposed the original idea; Z.W. performed the experiments and wrote the manuscript; T.S. and K.H. revised the manuscript; T.S., K.H. and Y.Z. supervised this study; K.H., Y.Z. and X.Y. provided guided advice for deep learning model training. Z.W., T.S., K.H. and X.Y. contributed equally to this work. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Key Laboratory of Land Satellite Remote Sensing Application, Ministry of Natural Resources of the People’s Republic of China with grant number KLSMNR-202208; the China High-Resolution Earth Observation System with grant number 21-Y20B01-9001-19/22 and 21-Y20B01-9003-19/22.

Data Availability Statement

The landslide dataset that supports our research in this paper is openly available in github at https://github.com/zhaoqiuw/landslide_dataset, accessed on 18 October 2022.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Davies, T. Landslide hazards, risks, and disasters: Introduction. In Landslide Hazards, Risks, and Disasters; Elsevier: Amsterdam, The Netherlands, 2015; pp. 1–16. [Google Scholar]
  2. Baum, R.L.; Godt, J.W.; Savage, W.Z. Estimating the timing and location of shallow rainfall-induced landslides using a model for transient, unsaturated infiltration. J. Geophys. Res. Earth Surf. 2010, 115, F03013. [Google Scholar] [CrossRef]
  3. Camps-Valls, G.; Tuia, D.; Gómez-Chova, L.; Jiménez, S.; Malo, J. Remote sensing image processing. Synth. Lect. Image Video Multimed. Process. 2011, 5, 1–192. [Google Scholar]
  4. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
  5. Dey, V.; Zhang, Y.; Zhong, M. A Review on Image Segmentation Techniques with Remote Sensing Perspective; ISPRS IC VII Symposium: Vienna, Austria, 2010; Volume 38. [Google Scholar]
  6. Liu, D.; Li, Y.; Lin, J.; Li, H.; Wu, F. Deep learning-based video coding: A review and a case study. ACM Comput. Surv. (CSUR) 2020, 53, 1–35. [Google Scholar] [CrossRef] [Green Version]
  7. Voulodimos, A.; Doulamis, N.; Doulamis, A.; Protopapadakis, E. Deep learning for computer vision: A brief review. Comput. Intell. Neurosci. 2018, 2018, 7068349. [Google Scholar] [CrossRef]
  8. Ghorbanzadeh, O.; Blaschke, T.; Gholamnia, K.; Meena, S.R.; Tiede, D.; Aryal, J. Evaluation of different machine learning methods and deep-learning convolutional neural networks for landslide detection. Remote Sens. 2019, 11, 196. [Google Scholar] [CrossRef] [Green Version]
  9. Badrinarayanan, V.; Kendall, A.; Cipolla, R. Segnet: A deep convolutional encoder-decoder architecture for image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 2481–2495. [Google Scholar] [CrossRef] [PubMed]
  10. Albawi, S.; Mohammed, T.A.; Al-Zawi, S. Understanding of a convolutional neural network. In Proceedings of the 2017 International Conference on Engineering and Technology (ICET), Antalya, Turkey, 21–23 August 2017; pp. 1–6. [Google Scholar]
  11. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 30. [Google Scholar] [CrossRef]
  12. Niu, Z.; Zhong, G.; Yu, H. A review on the attention mechanism of deep learning. Neurocomputing 2021, 452, 48–62. [Google Scholar] [CrossRef]
  13. Han, K.; Wang, Y.; Chen, H.; Chen, X.; Guo, J.; Liu, Z.; Tang, Y.; Xiao, A.; Xu, C.; Xu, Y.; et al. A survey on visual transformer. arXiv 2020, arXiv:2012.12556. [Google Scholar]
  14. Zhang, Z.; Zhang, X.; Peng, C.; Xue, X.; Sun, J. Exfuse: Enhancing feature fusion for semantic segmentation. In Proceedings of the European conference on computer vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 269–284. [Google Scholar]
  15. Hearst, M.A.; Dumais, S.T.; Osuna, E.; Platt, J.; Scholkopf, B. Support vector machines. IEEE Intell. Syst. Their Appl. 1998, 13, 18–28. [Google Scholar] [CrossRef] [Green Version]
  16. Dunteman, G.H. Principal Components Analysis; Sage: New York, NY, USA, 1989; p. 69. [Google Scholar]
  17. Shorten, C.; Khoshgoftaar, T.M. A survey on image data augmentation for deep learning. J. Big Data 2019, 6, 60. [Google Scholar] [CrossRef]
  18. Lateef, F.; Ruichek, Y. Survey on semantic segmentation using deep learning techniques. Neurocomputing 2019, 338, 321–348. [Google Scholar] [CrossRef]
  19. Fu, K.S.; Mui, J. A survey on image segmentation. Pattern Recognit. 1981, 13, 3–16. [Google Scholar] [CrossRef]
  20. Minaee, S.; Boykov, Y.Y.; Porikli, F.; Plaza, A.J.; Kehtarnavaz, N.; Terzopoulos, D. Image segmentation using deep learning: A survey. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 3523–3542. [Google Scholar] [CrossRef]
  21. Yi, L.; Zhijun, G. A review of segmentation method for MR image. In Proceedings of the 2010 International Conference on Image Analysis and Signal Processing, Zhejiang, China, 9–11 April 2010; pp. 351–357. [Google Scholar]
  22. Minaee, S.; Fotouhi, M.; Khalaj, B.H. A geometric approach to fully automatic chromosome segmentation. In Proceedings of the 2014 IEEE Signal Processing in Medicine and Biology Symposium (SPMB), Philadelphia, PA, USA, 13 December 2014; pp. 1–6. [Google Scholar]
  23. O’Shea, K.; Nash, R. An introduction to convolutional neural networks. arXiv 2015, arXiv:1511.08458. [Google Scholar]
  24. Svozil, D.; Kvasnicka, V.; Pospichal, J. Introduction to multi-layer feed-forward neural networks. Chemom. Intell. Lab. Syst. 1997, 39, 43–62. [Google Scholar] [CrossRef]
  25. Ramachandran, P.; Zoph, B.; Le, Q.V. Searching for activation functions. arXiv 2017, arXiv:1710.05941. [Google Scholar]
  26. Sun, M.; Song, Z.; Jiang, X.; Pan, J.; Pang, Y. Learning pooling for convolutional neural network. Neurocomputing 2017, 224, 96–104. [Google Scholar] [CrossRef]
  27. Basha, S.S.; Dubey, S.R.; Pulabaigari, V.; Mukherjee, S. Impact of fully connected layers on performance of convolutional neural networks for image classification. Neurocomputing 2020, 378, 112–119. [Google Scholar] [CrossRef] [Green Version]
  28. Muhammad, U.; Wang, W.; Chattha, S.P.; Ali, S. Pre-trained VGGNet architecture for remote-sensing image scene classification. In Proceedings of the 2018 24th International Conference on Pattern Recognition (ICPR), Beijing, China, 20–24 August 2018; pp. 1622–1627. [Google Scholar]
  29. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  30. Liu, Z.; Mao, H.; Wu, C.Y.; Feichtenhofer, C.; Darrell, T.; Xie, S. A convnet for the 2020s. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–20 June 2022; pp. 11976–11986. [Google Scholar]
  31. Zhou, Y.; Chen, S.; Wang, Y.; Huan, W. Review of research on lightweight convolutional neural networks. In Proceedings of the 2020 IEEE 5th Information Technology and Mechatronics Engineering Conference (ITOEC), Chongqing, China, 12–14 June 2020; pp. 1713–1720. [Google Scholar]
  32. Zhang, X.; Zhou, X.; Lin, M.; Sun, J. Shufflenet: An extremely efficient convolutional neural network for mobile devices. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 6848–6856. [Google Scholar]
  33. Howard, A.G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, H. Mobilenets: Efficient convolutional neural networks for mobile vision applications. arXiv 2017, arXiv:1704.04861. [Google Scholar]
  34. Tan, M.; Le, Q. Efficientnet: Rethinking model scaling for convolutional neural networks. In Proceedings of the International Conference on Machine Learning, PMLR, Long Beach, CA, USA, 9–15 June 2019; pp. 6105–6114. [Google Scholar]
  35. Chowdhary, K. Natural language processing. In Fundamentals of Artificial Intelligence; Springer: New Delhi, India, 2020; pp. 603–649. [Google Scholar]
  36. Mikolov, T.; Karafiát, M.; Burget, L.; Cernockỳ, J.; Khudanpur, S. Recurrent neural network based language model. In Proceedings of the 11th Annual Conference of the International-Speech-Communication-Association 2010, Makuhari, Japen, 1 January 2010; Volume 2, pp. 1045–1048. [Google Scholar]
  37. Chung, J.; Gulcehre, C.; Cho, K.; Bengio, Y. Empirical evaluation of gated recurrent neural networks on sequence modeling. arXiv 2014, arXiv:1412.3555. [Google Scholar]
  38. Lin, T.; Horne, B.G.; Tino, P.; Giles, C.L. Learning long-term dependencies in NARX recurrent neural networks. IEEE Trans. Neural Netw. 1996, 7, 1329–1338. [Google Scholar] [PubMed] [Green Version]
  39. Gehring, J.; Auli, M.; Grangier, D.; Yarats, D.; Dauphin, Y.N. Convolutional sequence to sequence learning. In Proceedings of the International Conference on Machine Learning, PMLR, Sydney, Australia, 6–11 August 2017; pp. 1243–1252. [Google Scholar]
  40. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  41. Chen, J.N.; Sun, S.; He, J.; Torr, P.H.; Yuille, A.; Bai, S. Transmix: Attend to mix for vision transformers. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–20 June 2022; pp. 12135–12144. [Google Scholar]
  42. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Virtual Event, 11–17 October 2021; pp. 10012–10022. [Google Scholar]
  43. Golubitsky, M.; Stewart, I. Recent advances in symmetric and network dynamics. Chaos Interdiscip. J. Nonlinear Sci. 2015, 25, 097612. [Google Scholar] [CrossRef] [PubMed]
  44. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In International Conference on Medical Image Computing and Computer-Assisted Intervention; Springer: Berlin/Heidelberg, Germany, 2015; pp. 234–241. [Google Scholar]
  45. Chaurasia, A.; Culurciello, E. Linknet: Exploiting encoder representations for efficient semantic segmentation. In Proceedings of the 2017 IEEE Visual Communications and Image Processing (VCIP), St. Petersburg, FL, USA, 10–13 December 2017; pp. 1–4. [Google Scholar]
  46. Lin, T.Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2117–2125. [Google Scholar]
  47. Zhao, H.; Shi, J.; Qi, X.; Wang, X.; Jia, J. Pyramid scene parsing network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2881–2890. [Google Scholar]
  48. Yu, F.; Koltun, V. Multi-scale context aggregation by dilated convolutions. arXiv 2015, arXiv:1511.07122. [Google Scholar]
  49. Chen, L.C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. Deeplab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected crfs. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 40, 834–848. [Google Scholar] [CrossRef] [PubMed]
  50. Chen, L.C.; Zhu, Y.; Papandreou, G.; Schroff, F.; Adam, H. Encoder-decoder with atrous separable convolution for semantic image segmentation. In Proceedings of the European conference on computer vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 801–818. [Google Scholar]
  51. Yu, W.; Luo, M.; Zhou, P.; Si, C.; Zhou, Y.; Wang, X.; Feng, J.; Yan, S. Metaformer is actually what you need for vision. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–20 June 2022; pp. 10819–10829. [Google Scholar]
  52. Bailer, C.; Varanasi, K.; Stricker, D. CNN-based patch matching for optical flow with thresholded hinge embedding loss. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 3250–3259. [Google Scholar]
  53. Ioffe, S.; Szegedy, C. Batch normalization: Accelerating deep network training by reducing internal covariate shift. In Proceedings of the International Conference on Machine Learning, PMLR, Lille, France, 7–9 July 2015; pp. 448–456. [Google Scholar]
  54. Ba, J.L.; Kiros, J.R.; Hinton, G.E. Layer normalization. arXiv 2016, arXiv:1607.06450. [Google Scholar]
  55. Borji, A.; Lin, S. SplitMixer: Fat Trimmed From MLP-like Models. arXiv 2022, arXiv:2207.10255. [Google Scholar]
  56. Milletari, F.; Navab, N.; Ahmadi, S.A. V-net: Fully convolutional neural networks for volumetric medical image segmentation. In Proceedings of the 2016 Fourth International Conference on 3D Vision (3DV), Stanford, CA, USA, 25–28 October 2016; pp. 565–571. [Google Scholar]
  57. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L.C. Mobilenetv2: Inverted residuals and linear bottlenecks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 4510–4520. [Google Scholar]
Figure 1. The source and production process of the landslide dataset.
Figure 1. The source and production process of the landslide dataset.
Sustainability 14 16311 g001
Figure 2. Data-augmented visualization results with color dithering and rotation transformation.
Figure 2. Data-augmented visualization results with color dithering and rotation transformation.
Sustainability 14 16311 g002
Figure 3. The encoder-decoder architecture.
Figure 3. The encoder-decoder architecture.
Sustainability 14 16311 g003
Figure 4. The framework of the proposed SCANet, which consists of the Poolformer as encoder, Separable Channel Attention Feature Pyramid Network (SCA-FPN) as decoder. S means Poolformer block.
Figure 4. The framework of the proposed SCANet, which consists of the Poolformer as encoder, Separable Channel Attention Feature Pyramid Network (SCA-FPN) as decoder. S means Poolformer block.
Sustainability 14 16311 g004
Figure 5. Metaformer block. Replace TokenMixer with attention to obtain Transformer block. Replace TokenMixer with pooling to obtain Poolformer block. Replace TokenMixer with spatial MLP to obtain MLP-like models block.
Figure 5. Metaformer block. Replace TokenMixer with attention to obtain Transformer block. Replace TokenMixer with pooling to obtain Poolformer block. Replace TokenMixer with spatial MLP to obtain MLP-like models block.
Sustainability 14 16311 g005
Figure 6. Separable Channel Attention Module.
Figure 6. Separable Channel Attention Module.
Sustainability 14 16311 g006
Figure 7. Visual analysis of evaluation metrics. (a) Visualization of the proposed SCANet compared with other networks (MobileNetv2-DeepLabv3Plus, MobileNetv2-UNet, MobileNetv2-FPN, MobileNetv2-PSPNet and MobileNetv2-LinkNet) using Mobilenet_v2 as encoder; (b) Visualization of the proposed SCANet compared with other networks (ResNet50-DeepLabv3Plus, ResNet50-UNet, ResNet50-FPN, ResNet50-PSPNet and ResNet50-LinkNet) using ResNet50 as encoder.
Figure 7. Visual analysis of evaluation metrics. (a) Visualization of the proposed SCANet compared with other networks (MobileNetv2-DeepLabv3Plus, MobileNetv2-UNet, MobileNetv2-FPN, MobileNetv2-PSPNet and MobileNetv2-LinkNet) using Mobilenet_v2 as encoder; (b) Visualization of the proposed SCANet compared with other networks (ResNet50-DeepLabv3Plus, ResNet50-UNet, ResNet50-FPN, ResNet50-PSPNet and ResNet50-LinkNet) using ResNet50 as encoder.
Sustainability 14 16311 g007
Figure 8. The visualization results of our proposed SCANet and mainstream semantic segmentation networks (MobileNetv2-DeepLabv3Plus, MobileNetv2-UNet, MobileNetv2-FPN, MobileNetv2-PSPNet and MobileNetv2-LinkNet) using Mobilenet_v2 as encoder.
Figure 8. The visualization results of our proposed SCANet and mainstream semantic segmentation networks (MobileNetv2-DeepLabv3Plus, MobileNetv2-UNet, MobileNetv2-FPN, MobileNetv2-PSPNet and MobileNetv2-LinkNet) using Mobilenet_v2 as encoder.
Sustainability 14 16311 g008
Figure 9. The visualization results of our proposed SCANet and mainstream semantic segmentation networks (ResNet50-DeepLabv3Plus, ResNet50-UNet, ResNet50-FPN, ResNet50-PSPNet and ResNet50-LinkNet) using ResNet50 as encoder.
Figure 9. The visualization results of our proposed SCANet and mainstream semantic segmentation networks (ResNet50-DeepLabv3Plus, ResNet50-UNet, ResNet50-FPN, ResNet50-PSPNet and ResNet50-LinkNet) using ResNet50 as encoder.
Sustainability 14 16311 g009
Figure 10. The heatmap that derived from the features before the network classification layer on different networks.
Figure 10. The heatmap that derived from the features before the network classification layer on different networks.
Sustainability 14 16311 g010
Table 1. The specification of the landslide remote sensing images dataset.
Table 1. The specification of the landslide remote sensing images dataset.
Landslide DatasetNumberSizeAugmentation
Training Set1802256 × 256
Validation/Test Set591256 × 256-
Table 2. Experimental settings.
Table 2. Experimental settings.
ConfigurationContents
Operating systemUbuntu 18.04.5 LTS
GPUNVIDIA GeForce RTX 3060 (12 GB RAM)
Deep-learning frameworkPytorch 1.11.0 and Torchvision 0.12.0
Parallel computer platformCuda 11.3 and Cudnn 8.2.0
ProgramPython 3.7.11
OptimizerSGDM
Learning rate0.001
LR policyPoly
Batch size16
Total epochs300
Momentum0.9
Data augmentationRandom flip and Random rotate
Loss functionCrossEntropy and Dice
Table 3. The quantitative results of mainstream semantic segmentation methods and the proposed SCANet. The best results are highlighted in bold, and the second-best results are underlined.
Table 3. The quantitative results of mainstream semantic segmentation methods and the proposed SCANet. The best results are highlighted in bold, and the second-best results are underlined.
MethodRecall (%)Precision (%)OA (%)F1-score (%)IoU (%)Params (M)Times (ms/Img)
EncoderDecoder
Mobilenet_v2DeepLabV3Plus88.4588.7695.0288.6079.544.49.388
Mobilenet_v2Unet88.8589.7195.3289.2880.646.68.217
Mobilenet_v2FPN88.7888.3194.9788.5479.444.28.243
Mobilenet_v2PSPNet84.0980.8692.1682.4470.132.37.245
Mobilenet_v2Linknet87.9988.6294.8988.3179.064.37.837
ResNet50DeepLabV3Plus89.8488.7795.2989.3180.6826.711.699
ResNet50Unet89.9690.2495.6790.1081.9832.511.955
ResNet50FPN90.9886.5894.9488.7379.7426.110.773
ResNet50PSPNet86.9185.2693.8482.0575.5724.39.325
ResNet50Linknet89.4890.1395.5589.8081.5031.212.624
PoolformerSCA-FPN90.8390.9896.0290.9183.9321.611.426
Table 4. The quantitative results of different semantic segmentation networks using ResNet50 and Poolformer as encoders, respectively. The results of semantic segmentation network using Poolformer as encoder are highlighted in bold.
Table 4. The quantitative results of different semantic segmentation networks using ResNet50 and Poolformer as encoders, respectively. The results of semantic segmentation network using Poolformer as encoder are highlighted in bold.
DecoderEncoderIoU (%)F1-score (%)
ResNet50Poolformer
Unet-81.9890.10
-84.7791.76
FPN-79.7488.73
-81.7189.94
PSPNet-75.5782.05
-76.3082.87
Linknet-81.5089.80
-82.7490.56
SCA-FPN-81.0288.89
-83.9390.91
Table 5. The quantitative results of different semantic segmentation networks using FPN and SCA-FPN as decoders, respectively.
Table 5. The quantitative results of different semantic segmentation networks using FPN and SCA-FPN as decoders, respectively.
EncoderDecoderIoU (%)F1-score (%)
ResNet50PoolformerFPNSCA-FPN
--79.7488.73
--81.0288.89
--81.7189.94
--83.9390.91
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Wang, Z.; Sun, T.; Hu, K.; Zhang, Y.; Yu, X.; Li, Y. A Deep Learning Semantic Segmentation Method for Landslide Scene Based on Transformer Architecture. Sustainability 2022, 14, 16311. https://doi.org/10.3390/su142316311

AMA Style

Wang Z, Sun T, Hu K, Zhang Y, Yu X, Li Y. A Deep Learning Semantic Segmentation Method for Landslide Scene Based on Transformer Architecture. Sustainability. 2022; 14(23):16311. https://doi.org/10.3390/su142316311

Chicago/Turabian Style

Wang, Zhaoqiu, Tao Sun, Kun Hu, Yueting Zhang, Xiaqiong Yu, and Ying Li. 2022. "A Deep Learning Semantic Segmentation Method for Landslide Scene Based on Transformer Architecture" Sustainability 14, no. 23: 16311. https://doi.org/10.3390/su142316311

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop