Next Article in Journal
Reconstruction of the Temporal Correlation Network of All-Cause Mortality Fluctuation across Italian Regions: The Importance of Temperature and Among-Nodes Flux
Previous Article in Journal
Adaptive Significance Levels in Tests for Linear Regression Models: The e-Value and P-Value Cases
Previous Article in Special Issue
Nighttime Image Stitching Method Based on Guided Filtering Enhancement
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

CSI-Former: Pay More Attention to Pose Estimation with WiFi

School of Computer Science and Technology, Nanjing Tech University, Nanjing 211816, China
*
Author to whom correspondence should be addressed.
Entropy 2023, 25(1), 20; https://doi.org/10.3390/e25010020
Submission received: 20 October 2022 / Revised: 11 December 2022 / Accepted: 18 December 2022 / Published: 22 December 2022
(This article belongs to the Special Issue Information Theory in Signal Processing and Image Processing)

Abstract

:
Cross-modal human pose estimation has a wide range of applications. Traditional image-based pose estimation will not work well in poor light or darkness. Therefore, some sensors such as LiDAR or Radio Frequency (RF) signals are now using to estimate human pose. However, it limits the application that these methods require much high-priced professional equipment. To address these challenges, we propose a new WiFi-based pose estimation method. Based on the Channel State Information (CSI) of WiFi, a novel architecture CSI-former is proposed to innovatively realize the integration of the multi-head attention in the WiFi-based pose estimation network. To evaluate the performance of CSI-former, we establish a span-new dataset Wi-Pose. This dataset consists of 5 GHz WiFi CSI, the corresponding images, and skeleton point annotations. The experimental results on Wi-Pose demonstrate that CSI-former can significantly improve the performance in wireless pose estimation and achieve more remarkable performance over traditional image-based pose estimation. To better benefit future research on the WiFi-based pose estimation, Wi-Pose has been made publicly available.

1. Introduction

Pose estimation is a basic task of human behavior evaluation. It aims to analyze important joints of human bodies such as limbs and faces through a series of multimedia signals, and promote the recognition of human behaviors [1,2,3,4,5,6,7,8,9,10]. Nevertheless, traditional image-based pose estimation methods [11,12,13,14,15,16,17,18,19,20] usually do not work well under partial occlusion and poor lighting conditions. For example, the camera cannot collect clear and easy-to-estimate human images in glare interference or complete darkness. In addition, public concerns about privacy issues also limit the application of image-based pose estimation. Thus, it is urgent to find a cross-modal method to realize pose estimation when it is hard to capture human images.
In previous exploratory work, people have tried human pose estimation based on multiple signals. Adib et al. [21] try to use radio signals to locate human bodies. However, this method can only perform coarse-grained positioning analysis of human bodies, not fine-grained behavior analysis. Another example is the use of radiofrequency [22] signals for positioning and poses estimation of human bodies. However, this method requires the establishment of a special radio frequency signal equipment that is uncommon in daily life. Therefore, the high cost and the strict requirements for the installation environment of this method limit its application.
Compared with the above sensors, WiFi equipment has the advantages of lower costs and easy-to-layout. Therefore, using WiFi equipment for pose estimation is a good choice. Traditional WiF devices work under the wireless transmission standard protocol of the IEEE 802.11n [23]. Under the principle that human bodies consist of hierarchical organizations with different dielectric parameters [24,25,26], it receives changing WiFi data packet information through changing body poses in the WiFi signal field. Then, the received wireless information is parsed into Channel State Information (CSI) with 30 sub-carrier groups according to Orthogonal Frequency Division Multiplexing (OFDM) and serves as the original input of the network. In past research, people were constantly trying to build better neural network models to extract more significant feature information from CSI [27,28,29,30,31,32,33]. The work of Hao et al. [30] realized CSI-based gesture estimation. They tried to find those features that are more sensitive to gesture actions from CSI, and successfully established a well-performed gesture estimation model.
Motivated by the work of Hao et al. [30], we attempt to seek a reasonable method to extract pose features from the sub-carries that are more sensitive to most poses. However, the previous work [34] takes a 16-layers Resnet as the pose feature extractor and equally extracts the pose feature on the 30 sub-carriers of CSI. It indicates that the network cannot pay more attention to the information-rich sub-carries. Of course, this limits the performance of the network. Recently, Transformer [35] has shown impressive performance in various natural language processing [36,37] and computer vision tasks [38,39,40,41,42], due to its powerful multi-head attention. Following the attention mechanism of Transformer, a new Performer [43] is proposed to realize attention-based long sequence data analysis such as protein modeling. Benefit from its Fast Attention Via positive Orthogonal Random features approach (FAVOR+), Performer can linearly optimize traditional softmax attention calculation. It indicates that Performer has an excellent attention mechanism and better space utilization over traditional Transformer. Motivated by these advantages of Performer, we attempt to incorporate an attention mechanism to extract more hidden pose features from long sequence CSI. In addition, compared with traditional methods, we are committed to making the information-rich sub-carries receive more attention.
Consequently, this paper addresses the above problems and attempts to design a novel network for paying more attention to the information-rich sub-carries. To achieve this, a novel multi-head attention-based network CSI-former is proposed for WiFi-based pose estimation. The proposed CSI-former is built in a teacher–student manner: a teacher network based on Alphapose and a student network based on Performer [43] and convolutional neural network. The Alphapose-based teacher network is developed to generate skeleton point annotation of the input human images. The skeleton point annotation is used as the ground truth of the human poses to optimize the student network. In the student network, an attention-based Encoder-Decoder structure is proposed to extract pose feature from CSI and match it with the pose labels from the teacher network. Under the supervision of the teacher network, the student network keeps self-optimized learning until it realizes pure CSI-based pose estimation.
To demonstrate the performance of CSI-former, a reasonable and persuasive WiFi-based dataset is necessary. As the basis of establishing the dataset, it is significant for us to determine the WiFi frequency band. Most traditional CSI collection methods rely on a 2.4 GHz WiFi band. Nevertheless, the research of Yu et al. [44] demonstrates that compared with 2.4 GHz, the 5 GHz signals have strong anti-interference and fewer interference sources. Thus, we regard that 5 GHz WiFi is more suitable for a WiFi-based pose estimation network. Based on this, 12 volunteers were invited to perform indoor activities while collecting their images and CSI. Thus, a new WiFi-based pose estimation dataset named Wi-Pose that is composed of images, skeleton point annotations, and CSI was constructed. To the best of our knowledge, there is no such public dataset yet. The experimental results on Wi-Pose demonstrate that CSI-former achieves excellent performance on WiFi-based human pose estimation.
The contributions of this paper are summarized as follows:
1.
We propose a novel architecture CSI-former that is composed of the attention mechanism and traditional convolutional neural network for WiFi-based pose estimation. To the best of our knowledge, CSI-former firstly realizes the effective integration of the multi-head attention to the field of cross-modal human pose estimation and significantly improves the performance.
2.
Differ from most previous methods, we establish the WiFi-based human body pose estimation dataset via a 5 GHz wireless WiFi signal and successfully demonstrate the effectiveness of 5 GHz WiFi in cross-modal pose estimation.
3.
We successfully established the novel WiFi-based pose estimation dataset Wi-Pose that is composed of 12 different actions such as bending, circling, crouching, pulling, running, walking, waving, etc. In addition, to ensure the rationality and persuasiveness of Wi-Pose, the invited 12 volunteers have different heights and weights.
4.
The proposed CSI-former has been evaluated on Wi-Pose and compared with the traditional network. The results demonstrate that CSI-former achieves state-of-the-art performance in WiFi-based pose estimation.
In summary, we propose a WiFi-based pose estimation algorithm CSI-former in this paper to solve the performance defects and privacy problems of traditional image-based human pose estimation methods. Moreover, we present the attention mechanism-based network framework to improve the algorithm and achieve satisfactory performance. We also publish the new dataset Wi-Pose at https://github.com/NjtechCVLab/Wi-PoseDataset, accessed on 19 October 2022 to facilitate future research.
The rest of the article is organized as follows. We introduce the related work for human pose estimation in Section 2. Section 3 describes the proposed CSI-former in detail. In Section 4, we introduce the details and results of the experiment. Finally, we conclude in Section 5 by highlighting the innovation of our work and discussing the future work.

2. Related Work

2.1. Video-Based Human Pose Estimation

Liu et al. [45] proposed a two-stream convolutional neural network architecture with a spatiotemporal network. The network uses the camera to capture videos for human body pose estimation. Each video is divided into two parts: a space part for describing static information of the scene and the object, a time part for describing motion information of the object and the camera. However, since the access of the two-stream convolutional network to a temporal context is restricted, it is unsuitable for modeling long-time span structures. Videos are one dimension higher than images. Therefore, some scholars optimize the two-dimensional convolutional network into a three-dimensional convolutional network for video image processing tasks.
Ji et al. [46] uses a 3D convolution kernel to extract spatiotemporal features of video data and obtain motion information of the video stream. The model has achieved good performance in the application scenarios of human behavior estimation in airport surveillance videos. Nevertheless, it remains the disadvantage that the 3D convolutional network has a large amount of calculation, which affects the algorithm efficiency. In the work of Wang et al. [27,46], the Openpose model is used to estimate poses of human bodies in videos, and spatiotemporal maps of the key points of human bodies are constructed. The spatiotemporal map convolutional network (ST-GCN) aims to extract spatiotemporal features of human bodies’ key points from continuous video frames. Then, the features are used for video action classification to achieve better human motion estimation performance.
However, although these methods have achieved good results, they are still limited to pure image-based pose estimation. It indicates that these methods will not work well in poor light, so the application range is difficult to broaden.

2.2. Sensors-Based Human Pose Estimation

People are constantly exploring the cross-modal human body positioning. By the gyroscope sensor in the bracelet, we can monitor a person’s steps and heartbeat. Cui et al. [47] pointed out that they can detect the number of people and objects in a room by using ultra-wideband radar. The number of people can be predicted by comparing extracted effective information about the radar reflection wave on the obstacle with the sensor data for detecting human bodies.
Khan et al. [48] proposed a human activity estimation method based on acceleration sensors. In some cases, two identification schemes have been established. The first is a low-level solution, which uses statistical signal functions and artificial neural networks (ANN) to identify the current state of behavior, including static, transitional, or dynamic. The second is an advanced scheme, which uses autoregressive (AR) modeling of acceleration signals, and then combines AR coefficients with the signal amplitude area and tilt angle to form an enhanced feature vector. Then, through linear discriminant analysis and artificial neural network, the obtained feature vectors are processed to realize estimation of specific human activities.

2.3. WiFi-Based Human Pose Estimation

Wigest [28] performs pose estimation by analyzing rising and falling edges of the signal change represented by the signal strength received by WiFi in different actions. For a single access point and three access points, the estimation accuracy rates reach 87.5% and 96%, respectively. Compared with received signal strength indicator (RSSI), CSI is a fine-grained value at the physical layer. It provides channel estimation for each sub-carrier of each transmission link and reflects the multipath effect caused by small-scale fading and micro-motion. In the work of Wang et al. [29], the CSI-SPEED model proposed by the CARM system quantifies the relationship between CSI amplitude changes and human motion speed and provides a model basis for subsequent research. However, its behavior extraction algorithm is imperfect, and the time complexity of this estimation method is high.
WiHear [27] utilizes CSI changes caused by lip movements with a special directional antenna gain and introduces it into the contour of mouth movements by using the local multipath effect and wavelet packet transform. It can solve the problem of micro-movement detection and achieve a predefined range for words. The average detection accuracy of no more than six words spoken by a single person is 91%, but it is not ideal in signal noise reduction, so strong directional antennas can only be used to reduce noise and improve estimation accuracy. Similarly, Wang et al. [49] proposed a Wi-Alarm system, which ignores the data preprocessing process and uses the support vector machine (SVM) to directly extract the original CSI amplitude mean and variance from human perception as features. Although it can significantly save the calculation cost, the extracted features are not accurate enough for original CSI data to be susceptible to interference from the external environment. The CSI cannot be fully utilized under limited time-domain statistical features. It eventually leads to a system crash and makes estimation accuracy restricted.
Overall, like most WiFi-based human body estimation networks, these methods can only perform rough human body positioning or crowd counting [34,50,51,52,53,54,55,56], but still cannot achieve a fine-grained estimation of the entire human body pose.

3. Methodology

3.1. Overview

CSI-former aims to realize WiFi-based pose estimation via a teacher–student network: a teacher network estimates human pose in videos by Alphapose, a student network learns human pose from CSI by Performer and convolutional neural network.
Alphapose: Alphapose refers to a traditional two-step frame network for image-based human pose estimation. It achieves high-performance image-based pose estimation via the human detection framework YOlOv3 [57] and the innovative regional multi-person poses estimation framework RMPE [11]. Since Alphapose has shown excellent performance on many public datasets, it is suitable for the teacher network of CSI-former. Thus, Alphapose is used as the ground truth tagger of CSI-former to identify 18 skeleton key points (nose, neck, shoulders, elbows, wrists, hips, knees, ankles, eyes, ears) from human images. Finally, the output of Alphapose is used as the poses ground truth to train the student network.
CSI: Channel State Information (CSI) refers to wireless state information obtained via Orthogonal Frequency Division Multiplexing (OFDM). OFDM converts the high-speed serial data stream into 30 low-speed parallel sub-data streams by decomposing the original channel into 30 mutually orthogonal sub-channels and modulates these sub-data streams to the orthogonal sub-channels for propagation. Each sub-data stream is called a sub-carrier. The amplitude and phase information of all sub-carriers constitute CSI. On this basis, the open-source tools [23] can obtain CSI via characteristics of wireless multipath propagation and analyze its state changes to analyze the changes of the surrounding environment. It indicates that pure CSI can also realize the pose estimation of surrounding human movements.
The original CSI is captured by a three-antenna WiFi transmitter and a three-antenna receiver. The transmitter continuously broadcasts WiFi signals to the outside. When human bodies of different poses pass by, the receiver receives changed wireless signals and parses them into a tensor of m × 30 × 3 × 3 size. Where m represents the number of WiFi packets received, 30 represents the number of wireless sub-carriers, and 3 × 3 represents a 3 × 3 array composed of three transmitting antennas and three receiving antennas.
Unfortunately, it is impossible to annotate real human poses with pure CSI. Thus, we use a camera parallel to the WiFi transmitting antenna to capture human pose images. Then, the captured images are processed by Alphapose of the teacher network to generate annotation information of human poses. Finally, under the supervision of the teacher work, the student network learns pose estimation from CSI.

3.2. Attention-Guided DeNoising

Inevitably, it tends to generate noise due to environmental influences during the collection of CSI. Under the impact of noise, the performance of the CSI-based pose estimation is bound to be restricted without effective denoising methods. For exploring denoising methods, it is significant to analyze the original CSI. As shown in Figure 1a, it records the images of a volunteer that performs actions after standing still for about 1.5 s. Figure 1b records the corresponding CSI at the timestamp. Since the work of Wang et al. [58] demonstrates that the effect of noise on different sub-carriers is highly correlated, it can be seen in Figure 1b that the amplitudes of all sub-carriers are changing very similarly at the same time when the volunteer is stationary. Nevertheless, when the volunteer performs actions, different sub-carriers have inconsistent changes in amplitude. It indicates that these essential pose feature information in CSI cannot be concealed by noise.
Moreover, Wang et al. [58] also demonstrate that traditional low-pass filters or median filters tend to achieve less-than-satisfactory performance in CSI denoising. As shown in Figure 1, we use the Butterworth low-pass filter that has a sampling rate of 1000 samples per second, and the ten-point median filter to denoise the CSI of Figure 1b respectively. It can be seen that compared to the original CSI, the CSI after filtering becomes very smooth. In addition, even during volunteer activities, the amplitude change of CSI is almost negligible in every small range. It indicates that the hidden pose features of CSI are filtered too. Undoubtedly, it will bring difficulties to the feature extraction of the network.
Inspired by the fact that the pose features cannot be concealed by noise and traditional filters have a poor performance in CSI denoising. It is reasonable to design a new network that can greatly dilute the influence of noise by paying more attention to the pose features. To this end, we propose an attention-guided denoising method (ADN) by CSI-former. Specifically, during the network training, the proposed CSI-former allocates more attention of the network to the sub-carriers that are more sensitive to poses through the parameter updating of the multi-head attention allocation algorithm. As Figure 1e shows, CSI-former pays more attention to these most sensitive sub-carriers that have more pose features and ignores the other sub-carriers with fewer features but much noise. Through the effective distribution of the attention, more pose features are extracted and noise is diluted. Since the sub-carriers are orthogonal to each other and all contain pose features, CSI-former can efficiently allocate attention without losing information.
In addition to ADN, we use multi-frame CSI to align a single-frame image to further dilute the possible impact of noise. In the system settings, the sampling rate of CSI is 100 Hz and the camera’s imaging frequency is 20 Hz. Through the synchronization of timestamps, every five CSI frames are aligned to one image frame. That is, for human body pose annotation information in each image frame, there is a corresponding CSI tensor with the size of 5 × 30 × 3 × 3.

3.3. Teacher Network: Alphapose

As the teacher network of CSI-former, Alphapose includes a two-step framework. First, a human body detector is used to form a human body detection box via the input images. After that, a pose estimation network will estimate the pose in the box. The pose skeleton points are finally output as annotation of the teacher network to the student network.
The collected data includes pose images I t (·) and CSI C t (·), aligned by timestamp t ( 0 , m ) . The original pose images I t (·) are processed through the teacher network Alphapose to obtain pose annotation P t (·), which is a 3 × 18 matrix composed of 18 skeleton key points coordinates ( x , y ) and their confidence c:
P t ( x , y , c ) ( 3 × 18 ) = A l p h a p o s e ( I t ( · ) ) , t ( 0 , m ) .
With the original pose annotation P t (·) from the teacher network, the student network needs to realize regression learning of 18 skeleton points of the human body. However, many previous works have demonstrated that it is easy to overfit and lose generalization by simply returning to 18 skeleton points [59]. Therefore, it is necessary to add the skeleton-point adjacency matrix (SAM) as the regular term. As shown in Figure 2, SAM consists of a 3 × 18 × 18 matrix ( x i , j , y i , j , c i , j ), ( i , j [ 1 , 2 , 3 , 18 ] ), where ( x , y , c ) represents coordinates of skeleton points and theirs confidence. Thus, SAM is a matrix obtained by two-dimensional expansion of ( x i , y i , c i ), i [ 1 , 2 , 3 , 18 ] in which x and y generation rules are the same:
x i , j = x i x j , i j ; x i , i = j . c i , j = c i × c j , i j ; c i , i = j .
SAM enhances the generalization ability of the network via taking relative displacement between skeleton points as an additional constraint. However, it also greatly increases the number of parameters that the network needs to return, and most of the attention is still needs to be paid to the regression of SAM diagonal value. Therefore, the application of the attention mechanism can significantly enhance the performance of the entire network.
Inspired by the SAM, the original pose annotation P t (·) will be expanded to obtain a label matrix with a size of 3 × 18 × 18 as the poses ground truth, which is called G t (·):
G t ( x , y , c ) ( 3 × 18 × 18 ) = S A M ( P t ( · ) )
Apart from images, the parsed CSI C t (·) is a tensor with a size of 30 × 3 × 3. Since the image sampling rate and the CSI sampling rate is 20 Hz and 100 Hz respectively, every five C t (·) will be aligned with one G t (·) through alignment of the time stream. Thus, original WiFi data W t (·) with a size of 5 × 30 × 3 × 3 for the student network is obtained:
W t ( · ) ( f r a m e s × 30 × 3 × 3 ) = k = t t + 4 C k ( · ) , f r a m e s = 5

3.4. Student Network: CSI-Former

The student network is composed of Performer and Convolutional Neural Network. The multi-head attention mechanism makes the student network improve the ability to extract pose features from CSI while learning annotations from the teacher network. Therefore, the student network includes three parts: encoder, feature extractor, and decoder.
Encoder : The encoder is developed to encode the original input W t to adapt to feature extraction. First, the input W t R 5 × 30 × 3 × 3 data size is reshaped to R 150 × 3 × 3 , which makes it correspond to data dimension of the teacher network so that makes convenient for network learning:
W t ( · ) ( 5 × 30 × 3 × 3 ) W t ( · ) ( 150 × 3 × 3 )
Then, the encoder directly uses bilinear interpolation to perform preliminary up-sampling on it, and expand it to R 150 × 18 × 18 :
W t ( · ) ( 150 × 18 × 18 ) = B i l i n e a r ( W t ( · ) )
The encoder can not only magnify the features of original data for easy extraction but also make it well adapted to the size of 18 skeleton points annotation from the teacher network.
Attention - Based Feature Extractor : Powerful feature extractor can better extract the feature information in the encoded data. Traditional feature extractors are composed of convolutional neural networks or pure ResNet [60]. However, these architectures always perform the same analysis on all input data instead of paying more attention to the more useful information, which limits the performance of the network.
Unlike the traditional methods, CSI-former uses a multi-layer composite attention-based Performer as the feature extractor (PAFE). The PAFE can make the student network pay more attention to those information-rich CSI inputs to realize efficient feature extraction. In addition, considering that the network needs to pay more attention to diagonal elements while taking into account non-diagonal elements when learning SAM, the PAFE can help realize reasonable distribution of attention. After a series of parameter stripping comparison tests, a stack of 12-layers Performer was finally selected. After the encoding matrix W t R 150 × 18 × 18 was extracted by Performer, the feature matrix F t R 150 × 18 × 18 with the same size as the encoding matrix is output:
F t ( · ) ( 150 × 18 × 18 ) = P A F E ( W t ( · ) ( 150 × 18 × 18 ) )
Decoder : The decoder is developed to decode the extracted feature matrix F t (·) to match label information G t (·). To achieve this, CSI-former uses the convolutional neural network with a two-layers architecture to be the decoder. The feature matrix F t R 150 × 18 × 18 will go through a convolutional layer firstly. In the layer, the 3 × 3 convolution kernel is used to initially release characteristic information as F t R 32 × 18 × 18 , which is followed by the BatchNorm layer and the ReLu layer. Then, the decoder selects a 1 × 1 convolution kernel to completely releases characteristic information S t R 2 × 18 × 18 , which contains pose coordinate information estimated by the student network. Finally, L 2 norm loss calculation is performed under the supervision of the teacher network:
S t ( x , y ) ( 2 × 18 × 18 ) = D e c o d e r ( F t ( · ) ) L o s s = L 2 ( G t ( x , y , c ) ( 3 × 18 × 18 ) , S t ( x , y ) ( 2 × 18 × 18 ) )
Finally, the gradient backpropagation is used to continuously optimize the student network until CSI-former can independently estimate human poses. It should be noted that the trained CSI-former can estimate any individual pose without an entire pose sequence. That is, CSI-former is trained by every single pose rather than ranking different poses.

3.5. Loss Calculation

Since CSI-former implements regression learning, the L 2 norm loss function which is commonly used in regression learning is suitable for the network [61]. The loss function needs to calculate loss between the output S t ( · ) of the student network and the ground truth G t ( · ) . In addition, the confidence information c i of G t ( · ) indicates its relevance to real poses, so it is necessary to take it into account when defining loss function. Finally, the loss function is defined as follows:
L = i = 0 17 G c i S x i G x i 2 2 + S y i G y i 2 2 ,
where · 2 2 represents L 2 loss calculation; G x i , G y i and S x i , S y i represents the ground truth and model prediction value of the i-th pose respectively. G c i represents confidence of the i-th skeleton point.

4. Experiments

4.1. Data Collection

To establish a rational and persuasive data set under real scenarios, we invited 12 volunteers of different heights and weights to perform multi-action activities indoors while simultaneously using WiFi devices and cameras to capture their wireless status information and images. Each volunteer did 12 different actions (bend, circle, crouch, jump, pull, push, run, sit down, stand up, throw, walk, wave) under guidance. Each action has a period of five seconds and repeats ten times. Since we cannot control the sampling time to exactly 5 s, the actual duration of each action is between 5 and 6 s. Moreover, our camera sampling rate is set to 20 Hz, thus the number of images per action is as follows:
N m i n = 12 × 5 × 10 × 20 = 12000 , N m a x = 12 × 6 × 10 × 20 = 14400 ,
We aligned and split the collected videos and CSI through time stamps, and put each video frame and its corresponding CSI data in a formatted file. Therefore, there are about 12,000 to 14,400 data for each action. The data number of Wi-Pose is 166,600 and the specific number of each action is shown in Figure 3. After analysis and sorting, the Wi-Pose that is composed of human images, its corresponding CSI, and skeleton point annotations was finally established.
For data division in the experiment, 80% of the collected data were used for the training set and the remaining 20% for the testing set. The number of the training set and the testing set are 132,847 and 33,753, respectively.

4.2. Model Evaluation

For evaluating the model, the percentage of correct key points (PCK) is one of the most persuasive evaluation algorithms for human pose key point detection models. It evaluates the CSI-former by calculating the ratio of the normalized distance between 18 detected key points and their corresponding ground truth that is less than the set threshold. The reference of normalized distance is developed via the torso diameter as follows:
T D k = G k R S x G k L H x 2 + G k R S y G k L H y 2 2 ,
where T D k is the diameter of the k-th body’s torso, G k R S and G k L H are the ground truth of the k-th body’s right shoulder and left Hip coordinates respectively. The calculated Euclidean distance between these two points is approximated as torso diameter. The specific algorithm of PCK is as follows:
P C K k i @ a j = 1 N i = 1 N δ S k i G k i 2 2 T D k a j ,
where a j is the j-th threshold of the algorithm, and the above formula P C K k i @ a j represents the PCK value of the i-th skeleton key point of the k-th human pose under the threshold, i = [ 1 , 2 , , 18 ] . The value of δ · refers to a boolean value whose value is one when the inequality in parentheses is true and otherwise zero. S k i and G k i respectively represent the coordinates of the i-th joint point of the k-th person identified by the model and the ground truth of the joint point. After the normalization calculation and the discrimination of δ · , the model finally obtains the average predicted PCK of each skeleton point of the testing set.
The threshold of PCK is usually set between 5 and 50. It should be noted that different thresholds are the evaluation standard of the model at different scales. Therefore, in this paper, we set the thresholds to various values between 5 and 50 to demonstrate the performance of our model from different evaluation scales. The larger the threshold is, the wider the error margin of skeleton point estimation is allowed. Thus, increasing the threshold may lead to higher PCK. However, lower thresholds represent more strict evaluation criteria, meaning lower thresholds can better demonstrate the model’s performance.

4.3. Implementation Details

CSI-former is implemented by Pytorch 1.7 and optimized by Adam optimizer. The batch size and epochs-number are eight and 50, respectively. The initial learning rate is set to 0.005 and is halved every ten epochs in the first 20 epochs and halved every 15 epochs in the last 30 epochs. We usually choose the model weights of the epoch which has the minimum loss. The epoch with minimum loss is generally between 45 to 50. The specific training details of the loss and the PCK are shown in Figure 4.

4.4. Ablation Study

In this section, we perform a series of ablation studies to analyze the impact of Performer layers on the performance of CSI-former. Under the strictly same training condition, four, eight, and twelve Performer layers were developed respectively to the feature extractor of CSI-former. The final result shown in Table 1 indicates that CSI-former with more layers of Performer tends to achieve higher PCK@5 on almost all skeleton points estimation. It demonstrates that under the same conditions, the more layers of Performer, the better performance of CSI-former. Thus, CSI-former is finally proposed via 12 layers stacked attention-based Performer. To further investigate the performance of CSI-former, we calculate the PCK of each skeleton point estimated via CSI-former by different thresholds of five, ten, twenty, thirty, forty, fifty, respectively. As shown in Table 2, it indicates CSI-former achieves excellent performance on the estimation of human poses.

4.5. Model Comparison

CSI-former uses 12-layers superimposed Performer as the feature extractor of the network. In order to prove the effectiveness of the multi-head attention mechanism, we compared the effects of CSI-former with pose estimation network using 16-layers superimposed Resnet as the feature extractor. After training models under the same experimental condition, we evaluated the two models with the same testing set. As shown in Table 3, it records the PCK@5 of 12 skeleton points in the testing set that are estimated by CSI-former and Resnet respectively. It is obvious that CSI-former achieves more superior performance over the Resnet on the estimation of ten different skeleton points and achieves a 2.7% increase in average PCK. The experimental result demonstrates that the multi-head attention mechanism significantly improves the pose estimation performance of the network.
Additionally, Figure 5 separately records the difference between the PCK results by CSI-former and Resnet for 18 skeleton points of different actions. It can be seen that the estimation performance of skeleton points of most actions (such as bend, jump, and run) has been greatly improved by CSI-former over Resnet. It indicates that through the biased allocation of attention by ADN, CSI-former significantly improves the estimation performance of most poses at the cost of reducing the accuracy of some poses. In other words, CSI-former successfully pays more attention to those sub-carriers that are sensitive to most actions and pays less attention to these subcarriers that are sensitive to individual actions to improve the overall predictive ability of the model.

4.6. Experimental Result

To observe the performance of CSI-former more directly, we estimate the poses of the testing set via CSI-former. Then, the estimated poses of CSI-former are drawn on the original images and compared with the pose estimation of the Alphapose-based teacher network, as shown in Figure 6. It can be observed that CSI-former achieves better performance over Alphapose on some skeleton points estimation. Additionally, we record the poses of volunteers with different heights and weights estimated by CSI-former in Figure 7. The experimental results demonstrate that the attention mechanism can significantly improve the performance of the WiFi-based pose estimation network and CSI-former achieves a superior pose estimation performance.

4.7. Result Discussion

As shown in the above sections, we finalized CSI-former through a series of ablation studies. We compare CSI-former with existing Resnet-based networks in Table 3, and CSI-former achieves higher PCK. As seen in Figure 6 and Figure 7, CSI-former has achieved satisfactory performance on WiFi-based human pose estimation.
Nevertheless, due to the constraints of time cost and experimental environment, CSI-former is trained by 12 activities performed by 12 volunteers. It indicates that the pose estimation results may have deviations when CSI-former is applied to other complex activities. Thus, we will expand more poses in more environments to our dataset Wi-Pose in future research. Since there is no publicly available dataset for WiFi-based human pose estimation currently, we have published Wi-Pose to facilitate future research. Moreover, we will continue to study new algorithms to improve CSI-former and enable it to estimate multi-person poses.

5. Conclusions

In this paper, we propose a WiFi-based pose estimation network CSI-former. Compared to traditional image-based pose estimation methods, CSI-former overcomes the disadvantages of cameras that cannot work in the dark and glare by using WiFi. Moreover, the WiFi-based method can solve the privacy concerns in human pose estimation well.
We design the architecture of CSI-former by 12 layers of Performer with multi-head attention to make the network pay more attention to the information that includes more pose features. The experimental results in Table 3 show that the PCK@5 of CSI-former achieves 0.5505, which is higher than the existing Resnet-based method whose PCK@5 is 0.5231. It demonstrates that CSI-former has better pose estimation performance than the existing Resnet-based method.
In addition, we successfully establish a novel WiFi-based human pose estimation dataset Wi-Pose and have published Wi-Pose to promote future research. CSI-former achieves state-of-the-art performance on Wi-Pose. Our future work will focus on further improving the structure of CSI-former to obtain better WiFi-based pose estimation capabilities.

Author Contributions

Conceptualization, L.Z.; data curation, Y.Z.; funding acquisition, Y.L.; methodology, Y.Z. and L.Z.; supervision, F.H. and Y.L.; Validation, C.X. and L.Z.; writing—original draft preparation, Y.Z.; writing—reviewing and editing, A.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work is partially supported by the National Natural Science Foundation of China (Grant No. 62101245), China Postdoctoral Science Foundation (Grant No. 2019M661999), Natural Science Research of Jiangsu Higher Education Institutions of China (19KJB520009), and Future Network Scientific Research Fund Project (Grant No. FNSRFP-2021-YB-21).

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

We are hugely grateful to the possible anonymous reviewers for their careful, unbiased, and constructive suggestions with respect to the original manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhang, X.; Zhang, F. Differentiable Spatial Regression: A Novel Method for 3D Hand Pose Estimation. IEEE Trans. Multimed. 2020, 24, 166–176. [Google Scholar] [CrossRef]
  2. Insafutdinov, E.; Pishchulin, L.; Andres, B.; Andriluka, M.; Schiele, B. Deepercut: A deeper, stronger, and faster multi-person pose estimation model. In European Conference on Computer Vision; Springer: Cham, Switzerland, 2016; pp. 34–50. [Google Scholar]
  3. Yan, S.; Xiong, Y.; Lin, D. Spatial temporal graph convolutional networks for skeleton-based action recognition. In Proceedings of the AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018; Volume 32. [Google Scholar]
  4. Yang, W.; Ouyang, W.; Li, H.; Wang, X. End-to-end learning of deformable mixture of parts and deep convolutional neural networks for human pose estimation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 3073–3082. [Google Scholar]
  5. Yang, W.; Ouyang, W.; Wang, X.; Ren, J.; Li, H.; Wang, X. 3d human pose estimation in the wild by adversarial learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 5255–5264. [Google Scholar]
  6. Liu, K.; Gao, L.; Khan, N.M.; Qi, L.; Guan, L. A multi-stream graph convolutional networks-hidden conditional random field model for skeleton-based action recognition. IEEE Trans. Multimed. 2020, 23, 64–76. [Google Scholar] [CrossRef]
  7. Ma, X.; Sang, N.; Xiao, S.; Wang, X. Learning a Deep Regression Forest for Head Pose Estimation from a Single Depth Image. J. Circuits Syst. Comput. 2021, 30, 2150139. [Google Scholar] [CrossRef]
  8. Dai, H.; Shi, H.; Liu, W.; Wang, L.; Liu, Y.; Mei, T. FasterPose: A faster simple baseline for human pose estimation. ACM Trans. Multimed. Comput. Commun. Appl. (TOMM) 2022, 18, 1–16. [Google Scholar] [CrossRef]
  9. Zhu, X.; Yang, Q.; Zhao, L.; Dai, Z.; He, Z.; Rong, W.; Sun, J.; Liu, G. An Improved Tiered Head Pose Estimation Network with Self-Adjust Loss Function. Entropy 2022, 24, 974. [Google Scholar] [CrossRef] [PubMed]
  10. Zheng, Z.; Ma, H.; Yan, W.; Liu, H.; Yang, Z. Training Data Selection and Optimal Sensor Placement for Deep-Learning-Based Sparse Inertial Sensor Human Posture Reconstruction. Entropy 2021, 23, 588. [Google Scholar] [CrossRef]
  11. Fang, H.S.; Xie, S.; Tai, Y.W.; Lu, C. Rmpe: Regional multi-person pose estimation. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2334–2343. [Google Scholar]
  12. Cao, Z.; Simon, T.; Wei, S.E.; Sheikh, Y. Realtime multi-person 2d pose estimation using part affinity fields. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 7291–7299. [Google Scholar]
  13. Newell, A.; Huang, Z.; Deng, J. Associative embedding: End-to-end learning for joint detection and grouping. arXiv 2016, arXiv:1611.05424. [Google Scholar]
  14. Newell, A.; Yang, K.; Deng, J. Stacked hourglass networks for human pose estimation. In European Conference on Computer Vision; Springer: Cham, Switzerland, 2016; pp. 483–499. [Google Scholar]
  15. Pishchulin, L.; Insafutdinov, E.; Tang, S.; Andres, B.; Andriluka, M.; Gehler, P.V.; Schiele, B. Deepcut: Joint subset partition and labeling for multi person pose estimation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 4929–4937. [Google Scholar]
  16. Ruggero Ronchi, M.; Perona, P. Benchmarking and error diagnosis in multi-instance pose estimation. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 369–378. [Google Scholar]
  17. Song, J.; Wang, L.; Van Gool, L.; Hilliges, O. Thin-slicing network: A deep structured model for pose estimation in videos. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4220–4229. [Google Scholar]
  18. Chu, X.; Ouyang, W.; Li, H.; Wang, X. Structured feature learning for pose estimation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 4715–4723. [Google Scholar]
  19. Huang, W.L.; Hung, C.Y.; Lin, I.C. Confidence-based 6D Object Pose Estimation. IEEE Trans. Multimed. 2021, 24, 3025–3035. [Google Scholar] [CrossRef]
  20. Jiao, Y.; Chen, H.; Feng, R.; Chen, H.; Wu, S.; Yin, Y.; Liu, Z. GLPose: Global-Local Representation Learning for Human Pose Estimation. ACM Transa. Multimed. Comput. Commun. Appl. (TOMM) 2022, 18, 128. [Google Scholar] [CrossRef]
  21. Adib, F.; Kabelac, Z.; Katabi, D.; Miller, R.C. 3d tracking via body radio reflections. In Proceedings of the 11th USENIX Symposium on Networked Systems Design and Implementation (NSDI 14), Seattle, WA, USA, 2–4 April 2014; pp. 317–329. [Google Scholar]
  22. Zhao, M.; Li, T.; Abu Alsheikh, M.; Tian, Y.; Zhao, H.; Torralba, A.; Katabi, D. Through-wall human pose estimation using radio signals. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7356–7365. [Google Scholar]
  23. Halperin, D.; Hu, W.; Sheth, A.; Wetherall, D. Tool release: Gathering 802.11 n traces with channel state information. ACM SIGCOMM Comput. Commun. Rev. 2011, 41, 53. [Google Scholar] [CrossRef]
  24. Christ, A.; Samaras, T.; Klingenböck, A.; Kuster, N. Characterization of the electromagnetic near-field absorption in layered biological tissue in the frequency range from 30 MHz to 6000 MHz. Phys. Med. Biol. 2006, 51, 4951. [Google Scholar] [CrossRef]
  25. Wang, F.; Han, J.; Dai, Z.; Ding, H.; Huang, D. Wipin: Operation-free person identification using wifi signals. arXiv 2018, arXiv:1810.04106. [Google Scholar]
  26. Christ, A.; Klingenbock, A.; Samaras, T.; Goiceanu, C.; Kuster, N. The dependence of electromagnetic far-field absorption on body tissue composition in the frequency range from 300 MHz to 6 GHz. IEEE Trans. Microw. Theory Tech. 2006, 54, 2188–2195. [Google Scholar] [CrossRef]
  27. Wang, G.; Zou, Y.; Zhou, Z.; Wu, K.; Ni, L.M. We can hear you with Wi-Fi! IEEE Trans. Mob. Comput. 2016, 15, 2907–2920. [Google Scholar] [CrossRef]
  28. Abdelnasser, H.; Youssef, M.; Harras, K.A. Wigest: A ubiquitous wifi-based gesture recognition system. In Proceedings of the 2015 IEEE Conference on Computer Communications (INFOCOM), Kowloon, Hong Kong, 26 April–1 May 2015; pp. 1472–1480. [Google Scholar]
  29. Wang, W.; Liu, A.X.; Shahzad, M.; Ling, K.; Lu, S. Device-free human activity recognition using commercial WiFi devices. IEEE J. Sel. Areas Commun. 2017, 35, 1118–1131. [Google Scholar] [CrossRef]
  30. Hao, Z.; Duan, Y.; Dang, X.; Liu, Y.; Zhang, D. Wi-SL: Contactless Fine-Grained Gesture Recognition Uses Channel State Information. Sensors 2020, 20, 4025. [Google Scholar] [CrossRef]
  31. Liu, W.; Jia, M.; Deng, Z.; Qin, C. MHSA-EC: An Indoor Localization Algorithm Fusing the Multi-Head Self-Attention Mechanism and Effective CSI. Entropy 2022, 24, 599. [Google Scholar] [CrossRef]
  32. Sun, Q.; Zhao, H.; Wang, J.; Chen, W. Deep Learning-Based Joint CSI Feedback and Hybrid Precoding in FDD mmWave Massive MIMO Systems. Entropy 2022, 24, 441. [Google Scholar] [CrossRef]
  33. Liu, W.; Wang, X.; Deng, Z. CSI Amplitude Fingerprinting for Indoor Localization with Dictionary Learning. Entropy 2021, 23, 1164. [Google Scholar] [CrossRef]
  34. Wang, F.; Han, J.; Zhang, S.; He, X.; Huang, D. CSI-Net: Unified human body characterization and pose recognition. arXiv 2018, arXiv:1810.03064. [Google Scholar]
  35. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention is all you need. arXiv 2017, arXiv:1706.03762. [Google Scholar]
  36. Brown, T.B.; Mann, B.; Ryder, N.; Subbiah, M.; Kaplan, J.; Dhariwal, P.; Neelakantan, A.; Shyam, P.; Sastry, G.; Askell, A.; et al. Language models are few-shot learners. arXiv 2020, arXiv:2005.14165. [Google Scholar]
  37. Devlin, J.; Chang, M.W.; Lee, K.; Toutanova, K. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv 2018, arXiv:1810.04805. [Google Scholar]
  38. Carion, N.; Massa, F.; Synnaeve, G.; Usunier, N.; Kirillov, A.; Zagoruyko, S. End-to-end object detection with transformers. In European Conference on Computer Vision; Springer: Cham, Switzerland, 2020; pp. 213–229. [Google Scholar]
  39. Parmar, N.; Vaswani, A.; Uszkoreit, J.; Kaiser, L.; Shazeer, N.; Ku, A.; Tran, D. Image transformer. In Proceedings of the International Conference on Machine Learning, Stockholm, Sweden, 10–15 July 2018; pp. 4055–4064. [Google Scholar]
  40. Zhu, A.; Zheng, Z.; Huang, Y.; Wang, T.; Jin, J.; Hu, F.; Hua, G.; Snoussi, H. CACrowdGAN: Cascaded attentional generative adversarial network for crowd counting. IEEE Trans. Intell. Transp. Syst. 2021, 23, 8090–8102. [Google Scholar] [CrossRef]
  41. Geng, Y.; Lian, Y.; Yang, S.; Zhou, M.; Cao, J. Reliable part guided multiple level attention learning for person re-identification. J. Circuits Syst. Comput. 2021, 30, 2150246. [Google Scholar] [CrossRef]
  42. Liu, X.; Li, B.; Chen, X.; Zhang, H.; Zhan, S. Content-based attention network for person image generation. J. Circuits Syst. Comput. 2020, 29, 2050250. [Google Scholar] [CrossRef]
  43. Choromanski, K.; Likhosherstov, V.; Dohan, D.; Song, X.; Gane, A.; Sarlos, T.; Hawkins, P.; Davis, J.; Mohiuddin, A.; Kaiser, L.; et al. Rethinking attention with performers. arXiv 2020, arXiv:2009.14794. [Google Scholar]
  44. Yu, F.; Jiang, M.; Liang, J.; Qin, X.; Hu, M.; Peng, T.; Hu, X. Expansion RSS-based indoor localization using 5G WiFi signal. In Proceedings of the 2014 International Conference on Computational Intelligence and Communication Networks, Bhopal, India, 14–16 November 2014; pp. 510–514. [Google Scholar]
  45. Liu, T.; Dong, X.; Wang, Y.; Dai, X.; You, Q.; Luo, J. Double-layer conditional random fields model for human action recognition. Signal Process. Image Commun. 2020, 80, 115672. [Google Scholar] [CrossRef]
  46. Ji, S.; Xu, W.; Yang, M.; Yu, K. 3D convolutional neural networks for human action recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 35, 221–231. [Google Scholar] [CrossRef] [Green Version]
  47. Choi, J.W.; Kim, J.H.; Cho, S.H. A counting algorithm for multiple objects using an IR-UWB radar system. In Proceedings of the 2012 3rd IEEE International Conference on Network Infrastructure and Digital Content, Beijing, China, 21–23 September 2012; pp. 591–595. [Google Scholar]
  48. Khan, A.M.; Lee, Y.K.; Lee, S.Y.; Kim, T.S. A triaxial accelerometer-based physical-activity recognition via augmented-signal features and a hierarchical recognizer. IEEE Trans. Inf. Technol. Biomed. 2010, 14, 1166–1172. [Google Scholar] [CrossRef]
  49. Wang, T.; Yang, D.; Zhang, S.; Wu, Y.; Xu, S. Wi-Alarm: Low-cost passive intrusion detection using WiFi. Sensors 2019, 19, 2335. [Google Scholar] [CrossRef] [Green Version]
  50. Wang, W.; Liu, A.X.; Shahzad, M. Gait recognition using wifi signals. In Proceedings of the 2016 ACM International Joint Conference on Pervasive and Ubiquitous Computing, Heidelberg, Germany, 12–16 September 2016; pp. 363–373. [Google Scholar]
  51. Zeng, Y.; Pathak, P.H.; Mohapatra, P. WiWho: WiFi-based person identification in smart spaces. In Proceedings of the 2016 15th ACM/IEEE International Conference on Information Processing in Sensor Networks (IPSN), Vienna, Austria, 11–14 April 2016; pp. 1–12. [Google Scholar]
  52. Zhang, J.; Wei, B.; Hu, W.; Kanhere, S.S. Wifi-id: Human identification using wifi signal. In Proceedings of the 2016 International Conference on Distributed Computing in Sensor Systems (DCOSS), Washington, DC, USA, 26–28 May 2016; pp. 75–82. [Google Scholar]
  53. Xin, T.; Guo, B.; Wang, Z.; Li, M.; Yu, Z.; Zhou, X. Freesense: Indoor human identification with Wi-Fi signals. In Proceedings of the 2016 IEEE Global Communications Conference (GLOBECOM), Washington, DC, USA, 4–8 December 2016; pp. 1–7. [Google Scholar]
  54. Lv, J.; Yang, W.; Man, D.; Du, X.; Yu, M.; Guizani, M. Wii: Device-free passive identity identification via WiFi signals. In Proceedings of the GLOBECOM 2017-2017 IEEE Global Communications Conference, Singapore, 4–8 December 2017; pp. 1–6. [Google Scholar]
  55. Xu, Q.; Chen, Y.; Wang, B.; Liu, K.R. Radio biometrics: Human recognition through a wall. IEEE Trans. Inf. Forensics Secur. 2017, 12, 1141–1155. [Google Scholar] [CrossRef]
  56. Yang, K.W.; Huang, Y.Y.; Huang, J.W.; Hsu, Y.R.; Wan, C.L.; Shuai, H.H.; Wang, L.C.; Cheng, W.H. Improving Crowd Density Estimation by Fusing Aerial Images and Radio Signals. ACM Trans. Multimed. Comput. Commun. Appl. (TOMM) 2022, 18, 1–23. [Google Scholar] [CrossRef]
  57. Redmon, J.; Farhadi, A. Yolov3: An incremental improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  58. Wang, W.; Liu, A.X.; Shahzad, M.; Ling, K.; Lu, S. Understanding and modeling of wifi signal based human activity recognition. In Proceedings of the 21st Annual International Conference on Mobile Computing and Networking, Paris, France, 7–11 September 2015; pp. 65–76. [Google Scholar]
  59. Pfister, T.; Charles, J.; Zisserman, A. Flowing convnets for human pose estimation in videos. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1913–1921. [Google Scholar]
  60. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  61. Daihong, J.; Sai, Z.; Lei, D.; Yueming, D. Multi-scale generative adversarial network for image super-resolution. Soft Comput. 2022, 26, 3631–3641. [Google Scholar] [CrossRef]
Figure 1. (a) records the images of a volunteer from stationary to running to stationary. (b) records the corresponding CSI of (a) in the time stamp. (c,d) represents the CSI after Butterworth low-pass filter and ten-pint median filter respectively. The highlighted parts of (e) are sub-carriers with richer features and more worthy of attention. The gray parts are sub-carriers that CSI-former tends to pay less attention to.
Figure 1. (a) records the images of a volunteer from stationary to running to stationary. (b) records the corresponding CSI of (a) in the time stamp. (c,d) represents the CSI after Butterworth low-pass filter and ten-pint median filter respectively. The highlighted parts of (e) are sub-carriers with richer features and more worthy of attention. The gray parts are sub-carriers that CSI-former tends to pay less attention to.
Entropy 25 00020 g001
Figure 2. The Alphapose-based teacher network is a two-step framework to analyze the input image and extract 18 skeleton point coordinates to form the SAM. The student network CSI-former is composed of the attention mechanism and convolutional neural network to extract a pose estimation matrix that is adapted to the SAM size from the CSI. Under the supervision of the teacher network, the student network continuously optimizes learning until it achieves the ability of CSI-based pose estimation.
Figure 2. The Alphapose-based teacher network is a two-step framework to analyze the input image and extract 18 skeleton point coordinates to form the SAM. The student network CSI-former is composed of the attention mechanism and convolutional neural network to extract a pose estimation matrix that is adapted to the SAM size from the CSI. Under the supervision of the teacher network, the student network continuously optimizes learning until it achieves the ability of CSI-based pose estimation.
Entropy 25 00020 g002
Figure 3. Wi-Pose includes 12 actions with a total of 166,600 data, and the amount of data for each action is roughly equal. We analyze and organize the collected raw data, and remove some data with unclear images or inaccurate skeleton points. Wi-Pose is finally composed of the remaining clear images, its corresponding CSI, and skeleton point annotation.
Figure 3. Wi-Pose includes 12 actions with a total of 166,600 data, and the amount of data for each action is roughly equal. We analyze and organize the collected raw data, and remove some data with unclear images or inaccurate skeleton points. Wi-Pose is finally composed of the remaining clear images, its corresponding CSI, and skeleton point annotation.
Entropy 25 00020 g003
Figure 4. CSI-former stops training when the loss converges and the PCK reaches the highest level. This goal can usually be achieved by training 50 epochs. Thus, CSI-former has been trained for a total of 50 epochs. The trend graph of loss and PCK during training is shown in (a) and (b) respectively.
Figure 4. CSI-former stops training when the loss converges and the PCK reaches the highest level. This goal can usually be achieved by training 50 epochs. Thus, CSI-former has been trained for a total of 50 epochs. The trend graph of loss and PCK during training is shown in (a) and (b) respectively.
Entropy 25 00020 g004
Figure 5. For each action, the vertical axis records the per-joint changes in the PCK@5 results of the proposed CSI-former relative to Resnet. The positive value means that CSI-former’s prediction accuracy for the skeleton joint is higher than Resnet.
Figure 5. For each action, the vertical axis records the per-joint changes in the PCK@5 results of the proposed CSI-former relative to Resnet. The positive value means that CSI-former’s prediction accuracy for the skeleton joint is higher than Resnet.
Entropy 25 00020 g005
Figure 6. (a) records the original human images of three volunteers. (b,c) records the pose estimation results of the three original human images by Alphapose and CSI-former, respectively. Each pose estimation draws 18 key skeleton points of the human body and connects these skeleton points in the order of the human body on the original image. It should be noted that the estimation of CSI-former on some skeleton points is better than Alphapose.
Figure 6. (a) records the original human images of three volunteers. (b,c) records the pose estimation results of the three original human images by Alphapose and CSI-former, respectively. Each pose estimation draws 18 key skeleton points of the human body and connects these skeleton points in the order of the human body on the original image. It should be noted that the estimation of CSI-former on some skeleton points is better than Alphapose.
Entropy 25 00020 g006
Figure 7. Wi-Pose includes different actions of persons with different heights and weights. (ac) are three groups of images of volunteers selected randomly. The first row of each group shows the real poses of the volunteers, and the corresponding second row shows the pose estimation results of CSI-former. The results demonstrate that CSI-former achieves superior performance on WiFi-based pose estimation.
Figure 7. Wi-Pose includes different actions of persons with different heights and weights. (ac) are three groups of images of volunteers selected randomly. The first row of each group shows the real poses of the volunteers, and the corresponding second row shows the pose estimation results of CSI-former. The results demonstrate that CSI-former achieves superior performance on WiFi-based pose estimation.
Entropy 25 00020 g007
Table 1. Comparisons between different Perormer layers in CSI-former. The bold value of each row is the optimal value.
Table 1. Comparisons between different Perormer layers in CSI-former. The bold value of each row is the optimal value.
Skeleton-Point PCK@5
Layers = 4Layers = 8Layers = 12
Nose0.48700.55980.5615
Neck0.51060.59150.5910
RShoulder0.49930.57470.5787
RElbow0.36680.43180.4387
RWrist0.27950.35040.3534
LShoulder0.45150.54550.5432
LElbow0.31710.43380.4484
LWrist0.28760.37420.4041
RHip0.49760.59550.6171
RKnee0.57850.66490.7093
RAnkle0.59710.71390.7422
LHip0.47000.58900.6050
LKnee0.54300.65090.6674
LAnkle0.52760.66190.6904
REye0.48720.56610.5918
LEye0.34400.50230.5240
REar0.49830.57490.6027
LEar0.21500.24560.2391
Average0.44210.53480.5505
Table 2. 12 Performer layers are used. The capitalized L and R refer to Left and Right, respectively. The bold value of each row is the optimal value.
Table 2. 12 Performer layers are used. The capitalized L and R refer to Left and Right, respectively. The bold value of each row is the optimal value.
Skeleton-PointPCK@5PCK@10PCK@20PCK@30PCK@40PCK@50
Nose0.56150.65430.74030.78570.81930.8421
Neck0.59100.68390.77060.81770.84810.8708
RShoulder0.57870.67690.76180.80970.84110.8663
RElbow0.43870.55830.67290.73530.77910.8120
RWrist0.35340.47090.59600.66600.71490.7547
LShoulder0.54320.65330.75830.81220.84360.8679
LElbow0.44840.56110.67560.74360.79080.8254
LWrist0.40410.51450.62470.69670.74620.7807
RHip0.61710.71190.80010.84520.87300.8920
RKnee0.70930.79050.86060.89300.91630.9322
RAnkle0.74220.81080.86060.89020.90600.9174
LHip0.60500.71180.79430.83980.86820.8881
LKnee0.66740.76440.83640.87100.89320.9080
LAnkle0.69040.75830.81660.84750.86850.7807
REye0.59180.67720.75500.80210.83160.8453
LEye0.52400.61880.71450.76870.80890.8357
REar0.60270.69020.76680.81080.84440.8643
LEar0.23910.29440.38740.46100.51690.5560
Average0.55050.64450.73290.78310.81720.8417
Table 3. Comparisons between attention-based and attention-free methods. The bold value of each row is the optimal value.
Table 3. Comparisons between attention-based and attention-free methods. The bold value of each row is the optimal value.
Skeleton-PointPCK@5
ResnetCSI-Former
Nose0.52710.5615
Neck0.55390.5910
RShoulder0.55250.5787
RElbow0.40990.4387
RWrist0.32340.3534
LShoulder0.49100.5432
LElbow0.38180.4484
LWrist0.35460.4041
RHip0.57600.6171
RKnee0.68800.7093
RAnkle0.72630.7422
LHip0.58750.6050
LKnee0.68500.6674
LAnkle0.67060.6904
REye0.57150.5918
LEye0.52730.5240
REar0.56200.6027
LEar0.22750.2391
Average0.52310.5505
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhou, Y.; Xu, C.; Zhao, L.; Zhu, A.; Hu, F.; Li, Y. CSI-Former: Pay More Attention to Pose Estimation with WiFi. Entropy 2023, 25, 20. https://doi.org/10.3390/e25010020

AMA Style

Zhou Y, Xu C, Zhao L, Zhu A, Hu F, Li Y. CSI-Former: Pay More Attention to Pose Estimation with WiFi. Entropy. 2023; 25(1):20. https://doi.org/10.3390/e25010020

Chicago/Turabian Style

Zhou, Yue, Caojie Xu, Lu Zhao, Aichun Zhu, Fangqiang Hu, and Yifeng Li. 2023. "CSI-Former: Pay More Attention to Pose Estimation with WiFi" Entropy 25, no. 1: 20. https://doi.org/10.3390/e25010020

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop