Next Article in Journal
Non-Enzymatically Colorimetric Bilirubin Sensing Based on the Catalytic Structure Disruption of Gold Nanocages
Previous Article in Journal
Remote Blood Pressure Estimation via the Spatiotemporal Mapping of Facial Videos
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Beam-Selection for 5G/B5G Networks Using Machine Learning: A Comparative Study

by
Efstratios Chatzoglou
1 and
Sotirios K. Goudos
2,*
1
Department of Computer Science, Hellenic Open University, Aristotelous 18, 26335 Patra, Greece
2
ELEDIA@AUTH, Department of Physics, Aristotle University of Thessaloniki, 54124 Thessaloniki, Greece
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(6), 2967; https://doi.org/10.3390/s23062967
Submission received: 31 January 2023 / Revised: 1 March 2023 / Accepted: 8 March 2023 / Published: 9 March 2023
(This article belongs to the Section Sensor Networks)

Abstract

:
A challenging problem in millimeter wave (mmWave) communications for the fifth generation of cellular communications and beyond (5G/B5G) is the beam selection problem. This is due to severe attenuation and penetration losses that are inherent in the mmWave band. Thus, the beam selection problem for mmWave links in a vehicular scenario can be solved as an exhaustive search among all candidate beam pairs. However, this approach cannot be assuredly completed within short contact times. On the other hand, machine learning (ML) has the potential to significantly advance 5G/B5G technology, as evidenced by the growing complexity of constructing cellular networks. In this work, we perform a comparative study of using different ML methods to solve the beam selection problem. We use a common dataset for this scenario found in the literature. We increase the accuracy of these results by approximately 30%. Moreover, we extend the given dataset by producing additional synthetic data. We apply ensemble learning techniques and obtain results with about 94% accuracy. The novelty of our work lies in the fact that we improve the existing dataset by adding more synthetic data and by designing a custom ensemble learning method for the problem at hand.

1. Introduction

One of the 5G proposals involves using machine learning tools to enhance communications, but this requires a lot of data to test algorithms and assess system performance. In addition, millimeter wave (mmWave) measurements for research in 5G multiple input multiple output (MIMO) require expensive campaigns in elaborated outside areas [1]. Real data related to wave propagation is difficult to obtain because of the time needed to set up the environment and configure the equipment, as well as the cost associated with the purchase of specialized hardware and human resources. This is one of the biggest challenges for 5G data generation, apart from Blockchain [2]. In this situation, synthetic data is a workable substitute as long as the environments in which it is generated are scalable and have a relatively low cost.
The problem of vehicular-to-infrastructure (V2I) or more general vehicle-to-everything (V2X) communication has attracted a lot of attention from the literature in the context of current 5G networks and in future 6G networks [3].
It is well-known that communication in the mmWave band faces several difficulties because of high attenuation and penetration loss. These problems can be overcome by phased arrays with directional beamforming by concentrating RF energy at the receiver [4,5].
In vehicular systems, mmWave MIMO is a method of exchanging sensor data, according to [6]. The fact that mmWave, as originally intended for this application, necessitates the pointing of narrow beams at both the transmitter and receiver poses a significant challenge. One of the interesting problems in this context consists of selecting the ideal pair of beams for analog beamforming (BF), with fixed beam codebooks and antenna arrays on both the transmitter and the receiver [7]. The authors of [8] provide a comprehensive overview of ML techniques for mmWave BF deployment in future wireless networks.
The research in [7] provides a dataset of small-scale parameters obtained from ray-tracing (RT) simulations, in which the simulation’s time-related snapshots resemble video frames. In this work, we apply several ML methods to produce better results than those found in the literature. More specifically, we apply both shallow and deep learners. We produce additional synthetic data (oversample) with Synthetic Minority Over-sampling Technique (SMOTE) [9] and we fine-tune the ML models using the GridSearch algorithm. The obtained results outperform those from the literature.
The main contributions of this work are summarized as follows:
  • We provide an effective way to produce synthetic data for a common dataset found in the literature.
  • We propose an ensemble learning composed of base shallow learners method to obtain more accurate results with low complexity.
  • We show that an optimal beam pair can be obtained using our approach up to an accuracy of about 94%.
To the best of the authors’ knowledge, this is the first time that an ensemble of classifiers is applied to the specific problem for optimal beam pair selection in 5G/6G cellular communications. The rest of this paper is organized as follows. Section 2 presents the related work. The problem description and the preprocessing process are given in Section 3. The numerical results are presented in Section 5. Finally, we discuss the results in Section 6, while the conclusion is drawn in Section 7.

2. Related Work

There are several survey papers in the literature that discuss different aspects of ML application to future wireless networks. The topic of antenna design for future networks using ML is discussed by the authors in [10]. In contrast, the topic of ML applications in vehicular networks from the networking point of view is examined in [11]. In the same context, the authors provide a thorough overview of various machine learning techniques applied to communication, network, and security components in a vehicular network in the works [12,13]. In another recent study, the authors in [14] report the most recent developments in the area of deep learning-based physical layer methods to pave the way for emerging applications of 6G.
Our contribution focused on the RAYMOBTIME dataset [7]. The latter dataset contains different simulated scenarios with Simulation of Urban MObility (SUMO) [15], i.e., a simulated automotive traffic program. The current dataset has already been analyzed by the literature, due to the diversity of options it offers. However, simpler studies are quite absent, i.e., most studies use extensive machine learning models to achieve optimal results.
First, in 2019, the authors of [16] analyzed this dataset with Deep Neural Networks (DNNs). With the assistance of LIDAR and separating their analysis into four different machine learning models, they managed to obtain 91.2% accuracy. The LIDAR sensor is responsible for identifying traces of light in order to calculate the distance between two objects.
Next, in 2021, the authors in [17] focused again on LIDAR technology, i.e., they used a Federated Learning methodology with a Convolutional Neural Network (CNN) of six layers, ReLU and PReLU activators, and the Softmax as the output function.
After that, the authors in [18] used both supervised and unsupervised ML models in their analysis. In the case of unsupervised learning, the mean absolute error (MAE) was 6.15.
In 2022, the authors in [19] propose a beam selection model, that will be based on CNN. Specifically, the latter CNN model should contain six layers (2D) and a linear layer. The GPS data were used at this point, and then four more linear layers were added. All layers had the ReLU activator, except the last one, which used Softmax. Their results showed an accuracy of 96.9% for the top-10 accuracy value for both the S008 and S009 sets.
The following study cited in [5] used both GPS and LIDAR data to improve the prediction rate of beam selection on top-10 accuracy. This analysis managed to achieve a top-10 accuracy of 91.11% with S008.
Authors in [20] proposed a backdoor attack algorithm to exploit the Federated Learning (FL) based mmWave beam selection method. Precisely, their attack triggered obstacles at specific locations on the road. They designed two target tasks; (a) to force the model to output a beam in a desired direction after the attack was triggered, and (b) to cause the model to send a beam with low signal strength. After confirming their experiments, they say that the beam selection system based on FL does not completely protect user privacy against certain attacks. To this end, they proposed a new defense method that can invalidate backdoor attacks, i.e., a dynamic norm clipping method that dynamically adjusts the size of the server clips, along with a backdoor detection method to efficiently provide a reliable signal. In the second case, a federated noise titration algorithm was proposed to show how the global model reacts to inputs with different levels of noise.
In [5], the authors designed deep learning architectures that can predict a set of top-K beam pairs with the assistance of non-RF sensor data such as GPS, camera, and LiDAR, improving the prediction accuracy by 3.32 to 43.9%. They also proposed that a fusion network exhibits 20–22% improvement in top-10 accuracy when comparing their technique against other state-of-the-art ones. Additionally, they calculated an optimization problem to choose the set of K candidate beam pairs, which decreased the beam selection time by 95–96%.
The authors of [21] proposed an estimation approach for both conventional and Reconfigurable Intelligent Surfaces (RIS) assisted massive MIMO systems of FL-based models. Input and output data were combined for each communication link, instead of using different neural networks for each task. They also proved the convergence of FL and demonstrated its superior performance over centralized learning (CL).
The authors propose FusionNet in [22], a dual-input neural network to predict the optimal beam using both the sub-6 GHz channels. They also used a deep learning method to activate the mmWave band antennas. The FusionNet outperformed the traditional fully connected neural network and the existing sub-6 GHz strategies in terms of both prediction accuracy and achievable rate.
However, in none of the above studies did the authors generate synthetic data to increase the dataset size. Additionally, we have not been able to find a study in the literature that used an ensemble learning approach for this dataset. Therefore, the goals of this study are to both augment the current dataset and improve accuracy results by introducing ensemble learning techniques.

3. Preliminaries

The authors in [7] have produced the dataset by combining simulations from ray tracing (RT) commercial software with the Simulation of Urban MObility (SUMO) traffic simulator. The dataset is generated representing a location in Rosslyn, VA, USA. Beam-selection is configured as a classification task with the optimal beam pair index as the target output. The operation frequency was 60 GHz.
In this study, we use the S000 scenario, the sane one the authors had analyzed in their study [7]. Regarding the S000 scenario and the data preprocessing steps, we used the ones the same authors had implemented in their study [7], i.e., they are directly related to the beam selection problem and the choice the antenna has to make during the beamforming process. This means that for the sake of Quality-of-Service (QoS), the antenna has to predict the next position of the user and send the current beam in the proper direction, eventually improving the whole QoS experience for each user.

4. Performance Metrics

In the remainder of this paper, we use the following multiclass classification metrics: [23,24].
1.
Accuracy is defined as the fraction of correct predictions. It is expressed as
A c c u r a c y = 1 N n = 1 N 1 1 ( z ^ n = z n )
where N is the number of test classes, z ^ n denotes the ML model class predicted label, z n is the true class label, and 1 ( x ) is the indicator function.
2.
Top-k Accuracy can be considered as a generalization of accuracy. The differentiation is that a prediction is regarded as accurate if the true label is connected to one of the top k forecasted scores. This is defined mathematically as
Top - k accuracy ( z , f ^ ) = 1 N n = 0 N 1 m = 1 k 1 ( f ^ n , m = z n )
where f ^ n , m denotes the predicted class for the n-th sample that corresponds to the m-th largest predicted score.
3.
Precision, which in multiclass classification is defined as
P r e c i s i o n = c = 1 C T P c c = 1 C T P c + F P c
where C is the number of classes, T P c is the number of true positive, and F P c is the number of false positive for class c, respectively.
4.
Recall is expressed as
R e c a l l = c = 1 C T P c c = 1 C T P c + F N c
where C is the number of classes, T P c is the number of true positive, and F N c is the number of false negative for class c respectively.
5.
F1 which aggregates Precision and Recall under the concept of harmonic mean. This is defined as
F 1 = 2 × Precision × Recall Precision + Recall .
F1 can be considered as a weighted average between Precision and Recall.
6.
Area Under the Curve (AUC) is the measure of a classifier’s ability to differentiate between classes. AUC computes the area under the receiver operating characteristic (ROC) curve. This summarizes the curve information into a single number. ROC is a probability curve, whereas AUC is a measure or degree of separability. It indicates how well the model can differentiate between classes. The model performs better with a higher AUC. In multiclass classification problems, the One-vs-one algorithm is a common solution. This algorithm determines the average AUC for all possible pairwise class combinations. The definition of this multiclass AUC metric, which is weighted uniformly, is given mathematically by [25]
1 C ( C 1 ) i = 1 C m > i C ( AUC ( i | m ) + AUC ( m | i )
where AUC ( i | m ) defines the AUC with class i as the positive class and class m as the negative class. In general, it is AUC ( i | m ) AUC ( m | i ) ) in the multiclass case.
For Regression we will use the metric Mean Absolute Error (MAE), which is defined as [26]:
M A E = 1 n i = 1 n | y i y ¯ i |
where n denotes the number of test patterns, y are the real measured data and y ¯ the predicted ones of the i-th data record.

5. Numerical Results

The main dataset was the same as in [7] in order to compare results. To identify the best ML model, we analyzed both shallow and Deep learning models. The reason is that shallow ML models can achieve good results faster and without losing much time configuring each model, but they cannot solve very complex problems. On the other hand, deep learning models can solve complex problems, but they need proper time to experiment and identify the optimal Deep learning network that we can create, i.e., the number of nodes, the number of layers, the correct choice of optimizer for each layer, etc. [27]. Our analysis followed the steps below:
  • Initial shallow analysis using several different ML learners, using the ML models of the original work [7].
  • Data augmentation with SMOTEN, to increase the samples of each class to at least 10. After that, we could analyze the dataset with the 10k stratified CV method.
  • Optimal hyperparameters search using the GridSearch algorithm.
  • Deep Learning analysis with both classification and regression tasks.
The numerical results were run on a Windows 10 Pro machine with an AMD Ryzen 7 2700, 64 GB of RAM, and a GTX 1060 6 GB GPU. Note that only the DNN models used the GPU as processing power; the shallow analysis was made with the CPU. The Python language with common libraries was used for the execution of ML models. Specifically, we used sklearn and imbalanced-learn.

5.1. Shallow Learning

Ensemble learning methods are a popular framework for machine learning applications, according to [28]. These methods train combinations of base models or base learners, which can be decision trees, neural networks, or any other supervised learning technique. Both regression and classification problems have been solved using ensemble methods. These methods have gained popularity because they have outperformed other single-model methods in multiple instances. The base learners to be combined in ensemble learning may be of the same or different types.
As a combination technique, learners like Random Forest (RF) [29] use Bootstrap Aggregating (or Bagging) [30]. This method improves the performance of individual trees by employing a voting scheme between their predictions. Prior to voting, each tree is grown using a distinct subset of the available training data to achieve diversity among the trees and reap the benefits of averaging the predictions of the uniquely grown trees. On the other hand, a Voting Classifier trains multiple models using different selected algorithms and returns the classification result based on the majority vote.
Ensemble models are valuable options, and they regularly produce excellent outcomes. They are less likely to overfit the data because they train multiple models with different data subsets. Since there are more models confirming the classification is in the right direction, they can provide greater precision.
First, we made a basic comparison with the initial results of the authors of RAYMOBTIME presented in [7]. We apply the following learners Linear Support Vector Classifier (LinearSVC) [31], Adaptive Boosting (AdaBoost) [32], Decision Trees (DT) [33], Random Forest (RF) [29], Deep Neural Network (DNN) [34], Naive Bayes [35], Artificial Neural Network (ANN) [36], k-Nearest Neighbors (kNN) [37], Stochastic Gradient Descent (SGD) Classifier [38], light gradient-boosting machine (LightGBM) [39], Logistic Regression [31], Extra Trees (ET) [40]. Table 1 presents the results of our analysis. As can be observed, there are small differences (2 to 5%) between the past and the current analysis. Even with other classification ML models, like Extra Trees, the results are not very accurate. The results indicate that our analysis is correct and very close to that of [7], as presented in the same table.
After further examination of the provided samples, it was observed that five classes had only one sample, and while the dataset was split into training and testing sets, these two sets did not contain the same number of samples. For example, a class had only one sample, and it was only on the training set. This made the evaluation phase of each ML model hard to predict because the models had to predict samples they did not know. For this reason, we removed these five classes (5, 10, 22, 38, and 54), merge the training and testing sets into one dataset, and split it again with a 70/30% analogy. It should be noted that by removing these five classes, the number of classes and the number of samples were reduced by 10% and by 0.01%, respectively. Since the analysis is based on an uneven set of data, we will focus on the AUC score to figure out how well we can predict. However, for the sake of completeness, the accuracy, and top-k metrics will also be present if they are applicable. We executed a new analysis with the same ML algorithms, but with 56 classes, as shown in Table 2. As can be observed, almost all ML models surpass the 50% threshold of accuracy by removing these classes. Specifically, only Naive Bayes presented worse results, in comparison to the 61 classes of analysis. About 90% of the ML models presented better results with 56 classes. On top of that, the results were improved by 40% in some cases, such as on DT.

5.2. Data Augmentation

To further improve the results of each classification analysis, a stratified k-fold cross-validation analysis should be made. To achieve this, each class must have at least 10 samples, to be able for the dataset to be split into 10 test sets. A dataset is considered to be imbalanced if the classes are not approximately equally represented. Thus, the specific dataset is imbalanced and we need to apply a data augmentation algorithm to correct this problem. For this reason, we acquired the SMOTE technique to oversample the dataset. This technique can create new samples based on the classes, by implementing the kNN algorithm.
Specifically, we used the Synthetic Minority Over-sampling TEchnique-Nominal Continuous (SMOTE-NC) algorithm [9], which is the SMOTE variant for oversampling mixed datasets of continuous and nominal features, with k_neighbors equal to 3 and random_state equals to 42. Since the k_neighbors parameter was equal to 3, each class should have at least 3 samples. So, to avoid overfitting the dataset with unnecessary data, we used the following pseudocode code of listing in Algorithm 1.
Algorithm 1 Pseudocode for SMOTE-NC samples separation.
1:
while  i < n u m O f F e a t u r e s   do
2:
   calculate s u m of unique values per feature
3:
   if  s u m < 26  then
4:
     if  s u m < 4  then
5:
        remove this feature from the dataset
6:
        populate the values of this feature, to be more than 3
7:
        add this feature into a new subset
8:
        continue
9:
       end if
10:
     remove this feature from the dataset
11:
     add this feature into a new subset
12:
   end if
13:
end while
Specifically, the Algorithm 1 kept one specific class, as the majority one, in order for the remaining classes to reach this number of samples. Then, if a class had fewer than four samples, their samples were re-copied. For instance, if a class had two samples, it got four samples, which were the same. With this method, all classes had at least three samples, for the SMOTE-NC algorithm to be able to oversample the dataset. After the Algorithm 1 execution, the SMOTE-NC algorithm was executed only to the training set. As a result, an additional 1311 samples (0.04%) were added to 28 classes, increasing the original sample size of the training set from 28,712 to 29,858.

5.3. Optimal Hyperparameters Search

In order to obtain the best performance we need to find the optimal hyperparameters for each shallow algorithm that we mentioned in Section 5.1. First, we made an analysis with GridSearchCV, to identify the latter hyperparameters. Second, we analyzed the shallow algorithms with the hyperparameters that GridSearchCV found.

5.3.1. GridSearchCV

To find the optimal hyperparameters, each shallow algorithm was executed with GridSearchCV 10 times. By using the Data Augmentation technique in Section 5.2, we managed to have at least 10 samples in each class of the dataset. Before executing GridSearchCV, we merged the SMOTE-NC training set and the test set into one dataset. As a result, each training set had 37,947 samples and each test set had 4217 samples, for each of the CV splits. To avoid overfitting, we executed the SMOTE-NC algorithm before executing GridSearchCV. This means that each ML model examined different oversampling data.
Table 3 illustrates the results of each algorithm. We were able to improve the prediction rate in most cases by fine-tuning each ML algorithm. Especially with LightGBM, which increased the prediction rate of the AUC score by ≈45%, reaching the highest AUC score so far, i.e., 85.38%.

5.3.2. Stratified Cross-Validation Results

After completing the analysis of Section 5.2 and Section 5.3, a question remains: can these models achieve the same results with stratified 10-fold cross-validation? And what will be the prediction rate of more complicated ML models, such as ensemble learning models? To this end, we added four ensemble methods, i.e., Voting-4, Voting-8, Bagging-DT, and Bagging-RF. Voting-4 contains the four best ML models, based on their AUC score, i.e., DT, NN, kNN, and LightGBM, while Voting-8 contains all ML models except LinearSVC, SGDClassifier, and Logistic Regression. Both voting algorithms used the “soft” method to predict the testing set. In soft voting, each individual classifier provides a probability that a particular data point belongs to a specific target class. The predictions are summed and weighted according to the importance of the classifier. The winner is the target label with the highest sum of weighted probabilities. This study analyzes and tries to improve the accuracy of the beamforming selection, with the purpose of increasing the quality of service for each user. Because of this, even a small improvement in beamforming selection can be seen as a success. Soft voting allows the selection of the beam pair with the highest probability. For this reason, we added the top-k accuracy metric and used soft voting.
Additionally, Bagging-DT uses the DT as the base estimator, while the Bagging-RF model uses the RF. Both Bagging models use the following parameters: bootstrap is false, the number of max_samples is equal to 10,000, the number of n_estimators is 100, and the oob_score is false.
Based on these parameters, Table 4 contains the results of all algorithms along with the ensemble models. Note that this table also contains the remaining evaluation metrics, such as the Precision (Prec), Recall, and F1 score. As can be observed from this table, the results in the AUC score are improving by 10% in each algorithm. This means that the process we used worked to make each ML model better at making predictions. Only Logistic Regression showed a smaller increase in the prediction rate, i.e., 6%, which was based on the convergence of this algorithm. This means that this algorithm needed more iterations to be trained correctly, and as a result, did not obtain the optimal training phase. Finally, 8 of the 15 ML models presented had an AUC score of more than 90%, which is a significant improvement over the results in Table 1. Voting-4 had the best AUC score with 94.5%.

5.4. DNN Analysis

We performed both classification and regression analysis using DL models. For the former analysis, two different DNN models were used, i.e., Multi-layer Perceptrons (MLP) and Denoising Stacked Autoencoders. For the latter analysis, only the CNN model was tested. Each analysis was implemented with a stratified 10-fold cross-validation. To achieve this, the SMOTE-NC algorithm was used, as described in Section 5.2.

5.4.1. DNN Classification

Each DNN model uses Model Checkpoint and Early Stopping. The former is capable of saving the weights after each epoch, while the latter is responsible for stopping the training phase if a targeted metric stop improves. In this study, the latter metric is loss. If the loss metric does not improve for two consecutive epochs, the training phase is terminated and the DNN model is re-trained using the most recent model checkpoint weights (optimal state). Table 5 highlights the details of each DNN model.
Table 6 highlights the results of each DNN model. As can be seen, DNN models were not able to predict more accurately than the Voting-4 model. However, both results, especially the AUC score of MLP are quite high (93.07%). So, these results validate that the methodology followed can improve the prediction rate of any ML model.
Figure 1 and Figure 2 illustrate the graph of the accuracy and the lost metrics per epoch for the 1st fold cross-validation training phase. This shows that the DNN models avoid overfitting, by using the Model Checkpoint, Early Stopping, and Dropout techniques. For instance, in the case of Figure 1 after the 40th epoch, the training accuracy is the same as or lower than the validation accuracy. So, the model stopped the training phase to avoid overfitting and retrained back to the optimal weight values.

5.4.2. DNN Regression

Regarding the regression analysis, we first examined the provided model of the authors, as stated in the GitHub repository [41]. So, we analyzed this model with the authors’ dataset for 500 epochs. The result was a Mean Square Percentage Error (MAPE) score equal to 85.34%, which is quite high.
For this reason, we created a new CNN model that was analyzed with the methodology of Section 5.1 and Section 5.2. This CNN ML model had the input Keras layer function, which was assigned the value of the total columns (5750). Then, the model had three consecutive layers, (a) a Conv1D, (b) an AveragePooling1D, and (c) a Batch normalization layer. After these layers, the activation function was placed, i.e., ReLU. These layers were added three times, and then a Dense layer was placed with 100 nodes. After that, a Dropout layer of 0.8 value was added, leading to the Flatten layer. The last layer had a Dense layer with only one node since the output activator was the Linear function.
As with the DNN classification models of Section 5.4.1, the CNN had the same values for learning rate and momentum. Also, CNN used the same optimizer, i.e., Stochastic Gradient Descent (SGD). Additionally, both Model Checkpoint and Early Stopping techniques were implemented, following a similar methodology as with the DNN classification models. Table 7 demonstrates the parameters of the CNN ML model.
After analyzing CNN with stratified 10-fold cross-validation with 56 classes and adding the oversampling effect with SMOTE-NC, the average MAE was equal to 5.32 , with only 5 epochs needed for the training phase.

6. Discussion

The findings of our work suggest that it is quite possible to find the optimal beam pair in 5G networks with a quite high probability of 94%. Such a result is feasible with low complexity without using time-consuming DL methods but rather using an ensemble approach with four base learners.
On the other hand, our work has certain limitations. More specifically, three facts limit our work. First, the low number of samples, or training data, may cause problems during the deep learning training phase. Second, the dataset is synthetic. As a result, as with any synthetic dataset, our ML models may be skewed in a real-world scenario. Third, the dataset needs much data preprocessing work to convert it to a proper set that is going to be analyzed by an ML model.

7. Conclusions

We have presented a comparative study of ML methods for the beam selection problem in 5G/B5G networks. We have given a complete framework of actions to improve classification accuracy. The accuracy of our results rises to above 92% using different ML models as well as the SMOTE-NC algorithm for oversampling the imbalanced classes. This is about 30% higher than that of the original paper. Moreover, both DL models gave results with accuracy higher than 90%. However, the higher accuracy values were obtained using ensemble learning techniques, which was expected since these methods could produce better results than individual learners. Both voting classifiers obtained results with accuracy higher than 95%. In addition, our study produced better results without using GPS or LIDAR data. Such data could possibly further improve the accuracy of the beamforming prediction. Also, we mostly used common shallow and Deep Learning models, which can be further improved into more complicated ML models, with the purpose of increasing the accuracy of each ML model. Thus, this study could be improved in future work by using more advanced Deep Learning models and analyzing additional data, like GPS or LIDAR

Author Contributions

Conceptualization, E.C. and S.K.G.; methodology, E.C.; software, E.C.; validation, E.C.; formal analysis, E.C. and S.K.G.; investigation, E.C.; resources, E.C. and S.K.G.; writing—original draft preparation, E.C. and S.K.G.; writing—review and editing, E.C. and S.K.G.; visualization, E.C.; supervision, S.K.G.; project administration, S.K.G.; All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

All data are available within the manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
CNNConvolutional Neural Network
DNNDeep Learning Neural Networks
DTDecision Trees
ETExtra Trees
lrlearning_rate
MAEMean Absolute Error
MLMachine Learning
mlnmax_leaf_nodes
MLPMulti-layer Perceptrons
MAPEMean Square Percentage Error
mslmin_samples_leaf
mssmin_samples_split
NNNeural Net
QoSQuality of Service
ReLURectified Linear Unit
RFRandom Forest
SCCSparse Categorical Crossentropy
SGDStohastic Gradient Descent
vsvar_smoothing

References

  1. MacCartney, G.R.; Yan, H.; Sun, S.; Rappaport, T.S. A flexible wideband millimeter-wave channel sounder with local area and NLOS to LOS transition measurements. In Proceedings of the 2017 IEEE International Conference on Communications (ICC), Paris, France, 21–25 May 2017; pp. 1–7. [Google Scholar] [CrossRef] [Green Version]
  2. Ali, A.; Rahim, H.A.; Pasha, M.F.; Dowsley, R.; Masud, M.; Ali, J.; Baz, M. Security, Privacy, and Reliability in Digital Healthcare Systems Using Blockchain. Electronics 2021, 10, 2034. [Google Scholar] [CrossRef]
  3. Noor-A-Rahim, M.; Liu, Z.; Lee, H.; Khyam, M.O.; He, J.; Pesch, D.; Moessner, K.; Saad, W.; Poor, H.V. 6G for Vehicle-to-Everything (V2X) Communications: Enabling Technologies, Challenges, and Opportunities. Proc. IEEE 2022, 110, 712–734. [Google Scholar] [CrossRef]
  4. Roh, W.; Seol, J.Y.; Park, J.; Lee, B.; Lee, J.; Kim, Y.; Cho, J.; Cheun, K.; Aryanfar, F. Millimeter-wave beamforming as an enabling technology for 5G cellular communications: Theoretical feasibility and prototype results. IEEE Commun. Mag. 2014, 52, 106–113. [Google Scholar] [CrossRef]
  5. Salehi, B.; Reus-Muns, G.; Roy, D.; Wang, Z.; Jian, T.; Dy, J.; Ioannidis, S.; Chowdhury, K. Deep Learning on Multimodal Sensor Data at the Wireless Edge for Vehicular Network. IEEE Trans. Veh. Technol. 2022, 71, 7639–7655. [Google Scholar] [CrossRef]
  6. González-Prelcic, N.; Ali, A.; Va, V.; Heath, R.W. Millimeter-Wave Communication with Out-of-Band Information. IEEE Commun. Mag. 2017, 55, 140–146. [Google Scholar] [CrossRef] [Green Version]
  7. Klautau, A.; Batista, P.; González-Prelcic, N.; Wang, Y.; Heath, R.W. 5G MIMO Data for Machine Learning: Application to Beam-Selection Using Deep Learning. In Proceedings of the 2018 Information Theory and Applications Workshop (ITA), San Diego, CA, USA, 11–16 February 2018; pp. 1–9. [Google Scholar] [CrossRef]
  8. Elhalawany, B.; Hashima, S.; Hatano, K.; Wu, K.; Mohamed, E. Leveraging Machine Learning for Millimeter Wave Beamforming in Beyond 5G Networks. IEEE Syst. J. 2022, 16, 1739–1750. [Google Scholar] [CrossRef]
  9. Chawla, N.; Bowyer, K.; Hall, L.; Kegelmeyer, W. SMOTE: Synthetic minority over-sampling technique. J. Artif. Intell. Res. 2002, 16, 321–357. [Google Scholar] [CrossRef]
  10. Khan, M.; Hossain, S.; Mozumdar, P.; Akter, S.; Ashique, R. A review on machine learning and deep learning for various antenna design applications. Heliyon 2022, 8, e09317. [Google Scholar] [CrossRef] [PubMed]
  11. Tan, K.; Bremner, D.; Le Kernec, J.; Zhang, L.; Imran, M. Machine learning in vehicular networking: An overview. Digit. Commun. Netw. 2022, 8, 18–24. [Google Scholar] [CrossRef]
  12. Tang, F.; Mao, B.; Kato, N.; Gui, G. Comprehensive survey on machine learning in vehicular network: Technology, applications and challenges. IEEE Commun. Surv. Tutor. 2021, 23, 2027–2057. [Google Scholar] [CrossRef]
  13. Tang, F.; Kawamoto, Y.; Kato, N.; Liu, J. Future Intelligent and Secure Vehicular Network Toward 6G: Machine-Learning Approaches. Proc. IEEE 2020, 108, 292–307. [Google Scholar] [CrossRef]
  14. Ozpoyraz, B.; Dogukan, A.; Gevez, Y.; Altun, U.; Basar, E. Deep Learning-Aided 6G Wireless Networks: A Comprehensive Survey of Revolutionary PHY Architectures. IEEE Open J. Commun. Soc. 2022, 3, 1749–1809. [Google Scholar] [CrossRef]
  15. Eclipse. SUMO. Available online: https://www.eclipse.org/sumo/ (accessed on 14 August 2022).
  16. Dias, M.; Klautau, A.; González-Prelcic, N.; Heath, R.W. Position and LIDAR-Aided mmWave Beam Selection using Deep Learning. In Proceedings of the 2019 IEEE 20th International Workshop on Signal Processing Advances in Wireless Communications (SPAWC), Cannes, France, 2–5 July 2019; pp. 1–5. [Google Scholar] [CrossRef]
  17. Mashhadi, M.B.; Jankowski, M.; Tung, T.Y.; Kobus, S.; Gündüz, D. Federated mmWave Beam Selection Utilizing LIDAR Data. IEEE Wirel. Commun. Lett. 2021, 10, 2269–2273. [Google Scholar] [CrossRef]
  18. Mukhtar, H.; Erol-Kantarci, M. Machine Learning-Enabled Localization in 5G using LIDAR and RSS Data. In Proceedings of the 2021 IEEE Symposium on Computers and Communications (ISCC), Athens, Greece, 5–8 September 2021; pp. 1–6. [Google Scholar] [CrossRef]
  19. Zecchin, M.; Mashhadi, M.B.; Jankowski, M.; Gündüz, D.; Kountouris, M.; Gesbert, D. LIDAR and Position-Aided mmWave Beam Selection With Non-Local CNNs and Curriculum Training. IEEE Trans. Veh. Technol. 2022, 71, 2979–2990. [Google Scholar] [CrossRef]
  20. Zhang, Z.; Yang, R.; Zhang, X.; Li, C.; Huang, Y.; Yang, L. Backdoor Federated Learning-Based mmWave Beam Selection. IEEE Trans. Commun. 2022, 70, 6563–6578. [Google Scholar] [CrossRef]
  21. Elbir, A.; Coleri, S. Federated Learning for Channel Estimation in Conventional and RIS-Assisted Massive MIMO. IEEE Trans. Wirel. Commun. 2022, 21, 4255–4268. [Google Scholar] [CrossRef]
  22. Gao, F.; Lin, B.; Bian, C.; Zhou, T.; Qian, J.; Wang, H. FusionNet: Enhanced Beam Prediction for mmWave Communications Using Sub-6 GHz Channel and a Few Pilots. IEEE Trans. Commun. 2021, 69, 8488–8500. [Google Scholar] [CrossRef]
  23. Grandini, M.; Bagli, E.; Visani, G. Metrics for Multi-Class Classification: An Overview. arXiv 2020, arXiv:2008.05756. [Google Scholar] [CrossRef]
  24. Sokolova, M.; Lapalme, G. A systematic analysis of performance measures for classification tasks. Inf. Process. Manag. 2009, 45, 427–437. [Google Scholar] [CrossRef]
  25. Hand, D.; Till, R. A Simple Generalisation of the Area Under the ROC Curve for Multiple Class Classification Problems. Mach. Learn. 2001, 45, 171–186. [Google Scholar] [CrossRef]
  26. Ahmad, N.; Ghadi, Y.; Adnan, M.; Ali, M. Load Forecasting Techniques for Power System: Research Challenges and Survey. IEEE Access 2022, 10, 71054–71090. [Google Scholar] [CrossRef]
  27. Goodfellow, I.; Bengio, Y.; Courville, A. Deep Learning; MIT Press: Cambridge, MA, USA, 2016; Available online: http://www.deeplearningbook.org (accessed on 25 February 2023).
  28. Zhang, C.; Ma, Y. Ensemble Machine Learning: Methods and Applications; Springer Publishing Company: Berlin/Heidelberg, Germany, 2012. [Google Scholar]
  29. Breiman, L. Random Forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef] [Green Version]
  30. Breiman, L. Bagging Predictors. Mach. Learn. 1996, 24, 123–140. [Google Scholar] [CrossRef] [Green Version]
  31. Fan, R.E.; Chang, K.W.; Hsieh, C.J.; Wang, X.R.; Lin, C.J. LIBLINEAR: A Library for Large Linear Classification. J. Mach. Learn. Res. 2008, 9, 1871–1874. [Google Scholar]
  32. Freund, Y.; Schapire, R. A Decision-Theoretic Generalization of On-Line Learning and an Application to Boosting. J. Comput. Syst. Sci. 1997, 55, 119–139. [Google Scholar] [CrossRef] [Green Version]
  33. Moore II, D.H. Classification and regression trees, by Leo Breiman, Jerome H. Friedman, Richard A. Olshen, and Charles J. Stone. Brooks/Cole Publishing, Monterey, 1984,358 pages, $27.95. Cytometry 1987, 8, 534–535. [Google Scholar] [CrossRef]
  34. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
  35. Zhang, H. The Optimality of Naive Bayes. In Proceedings of the Seventeenth International Florida Artificial Intelligence Research Society Conference (FLAIRS 2004), Miami Beach, FL, USA, 17–19 May 2004; Barr, V., Markov, Z., Eds.; AAAI Press: Washington, DC, USA, 2004. [Google Scholar]
  36. Christodoulou, C.; Georgiopoulos, M. Applications of Neural Networks in Electromagnetics; Artech House: Norwood, MA, USA, 2001. [Google Scholar]
  37. Altman, N.S. An Introduction to Kernel and Nearest-Neighbor Nonparametric Regression. Am. Stat. 1992, 46, 175–185. [Google Scholar]
  38. Chatzoglou, E.; Kambourakis, G.; Kolias, C.; Smiliotopoulos, C. Pick Quality Over Quantity: Expert Feature Selection and Data Preprocessing for 802.11 Intrusion Detection Systems. IEEE Access 2022, 10, 64761–64784. [Google Scholar] [CrossRef]
  39. Ke, G.; Meng, Q.; Finley, T.; Wang, T.; Chen, W.; Ma, W.; Ye, Q.; Liu, T.Y. LightGBM: A Highly Efficient Gradient Boosting Decision Tree. In Advances in Neural Information Processing Systems 30; Guyon, I., Luxburg, U.V., Bengio, S., Wallach, H., Fergus, R., Vishwanathan, S., Garnett, R., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2017; pp. 3146–3154. [Google Scholar]
  40. Geurts, P.; Ernst, D.; Wehenkel, L. Extremely randomized trees. Mach. Learn. 2006, 63, 3–42. [Google Scholar] [CrossRef] [Green Version]
  41. Oliveira, A. 5gm-beam-selection. Available online: https://github.com/lasseufpa/5gm-beam-selection (accessed on 14 August 2022).
Figure 1. MLP accuracy per epoch (top) and lost per epoch (bottom) for the 1st fold training phase.
Figure 1. MLP accuracy per epoch (top) and lost per epoch (bottom) for the 1st fold training phase.
Sensors 23 02967 g001
Figure 2. Autoencoder accuracy per epoch (top) and lost per epoch (bottom) for the 1st fold training phase.
Figure 2. Autoencoder accuracy per epoch (top) and lost per epoch (bottom) for the 1st fold training phase.
Sensors 23 02967 g002
Table 1. Results comparison with the original work of [7], based on the script’s execution that the original authors published with [7]. With bold font are highlighted the optimal results of each column.
Table 1. Results comparison with the original work of [7], based on the script’s execution that the original authors published with [7]. With bold font are highlighted the optimal results of each column.
ClassifiersAccuracy (%) in [7]Accuracy (%) This Study
LinearSVC33.2035.62
AdaBoost55.0054.96
Decision Trees55.5052.06
Random Forest63.2057.88
DNN63.2062.20
Naive Bayes8.70
Artificial Neural Net (ANN)40.50
kNN43.04
SGDClassifier37.80
LightGBM26.66
Logistic Regression43.76
Extra Trees (ET)52.66
Table 2. Results based on the execution of each algorithm, but with 56 classes. With bold fold highlighted the optimal results of each column are.
Table 2. Results based on the execution of each algorithm, but with 56 classes. With bold fold highlighted the optimal results of each column are.
ClassifiersAUC (%)Accuracy (%)Top-2Top-5Top-10
Naive Bayes64.184.7910.4423.1129.44
Decision Trees84.8592.9494.3995.1695.60
Random Forest74.4587.4496.0999.1699.33
AdaBoost50.2060.2870.8186.3693.87
LinearSVC74.4279.25
NN67.0088.9695.8998.6499.18
kNN76.7387.6894.4497.3797.47
SGDClassifier70.4659.52
LightGBM49.9740.4951.2851.5252.40
Logistic Regression63.4976.8390.7397.2398.58
ET50.0058.5269.2385.9093.53
Table 3. Results with the GridSearchCV parameters analysis. Bold highlights best results.
Table 3. Results with the GridSearchCV parameters analysis. Bold highlights best results.
ClassifiersAUCAccuracyTop-2Top-5Top-10
DT83.6793.5994.6294.6494.75
Random Forest66.4986.9696.3499.4899.73
NN84.1292.3594.4499.0499.48
kNN76.7387.6894.4497.3797.47
Logistic Regression80.1685.3795.8399.2199.57
Naive Bayes79.1414.1230.9766.2084.15
AdaBoost50.2060.2870.7886.3593.87
LinearSVC76.5383.61
SGDClassifier72.6382.86
LightGBM85.3895.1098.3799.5299.71
ET68.7485.9596.2299.1799.69
Table 4. Stratified 10-fold cross-validation results. Bold highlights best results.
Table 4. Stratified 10-fold cross-validation results. Bold highlights best results.
ClassifierAUCAccuracyPrecRecallF1Top-2Top-5Top-10
DT93.7794.5689.0087.6987.5495.3795.4195.54
RF85.2287.9183.6770.8974.5796.6399.5299.80
NN93.1593.6789.6086.4987.0097.8999.4199.64
kNN90.6691.7190.0781.5784.4290.6696.3998.32
Logistic Regression91.3985.8790.2583.2485.5896.1999.3399.69
Naive Bayes79.6417.2240.8871.3844.7436.8868.3384.95
AdaBoost50.7458.713.423.182.5568.9984.3591.84
LinearSVC89.5584.489.8379.6483.12
SGDClassifier87.7283.3185.1176.0178.16
LightGBM93.9095.6492.0887.9389.1898.6199.5399.77
ET83.4185.8582.4667.3371.5495.9899.2499.74
Voting-494.5095.8691.8987.1389.6698.7999.5899.76
Voting-894.1295.4092.1088.3989.4798.2299.5999.87
Bagging-DT92.0694.0091.5784.3286.7898.7499.6099.68
Bagging-RF87.0789.5587.3074.5478.5497.9399.6499.81
Table 5. DNN classification parameters per model.
Table 5. DNN classification parameters per model.
ParametersMLPAutoencoder
ActivatorReLUReLU
Output activatorSoftmaxSoftmax
InitializerHe_uniform
OptimizerSGDSGD
Momentum0.90.9
Dropout0.15/0.10.1
Learning rate0.010.01
LossSCCSCC
Batch normalizationYesYes
Layers1116
Nodes (per layer)500/450/400/
350/300/250/
200/160/120/
60/56
300/250/200/
100/100/50/25/
10/25/50/100/
100/200/250/300/
56
Batch size3232
Table 6. DNN classification results per model. Bold highlights best results.
Table 6. DNN classification results per model. Bold highlights best results.
ModelAUCAccuracyPrecisionRecallF1Epochs
MLP93.0794.1885.9086.2985.1750.6
Autoencoder87.2891.6874.7874.7973.2762.5
Table 7. DNN regression model parameters.
Table 7. DNN regression model parameters.
CNN
ActivatorReLU
Output activatorLinear
OptimizerSGD
Momentum0.9
Dropout0.8
Learning rate0.01
LossMAE
Batch normalizationYes
Layers6
Nodes (per layer)Conv1D(128,10)/
Conv1D(64,12)/
Conv1D(32,10)/
Dense(100)/
Flatten()/
Dense(1)
Batch size32
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chatzoglou, E.; Goudos, S.K. Beam-Selection for 5G/B5G Networks Using Machine Learning: A Comparative Study. Sensors 2023, 23, 2967. https://doi.org/10.3390/s23062967

AMA Style

Chatzoglou E, Goudos SK. Beam-Selection for 5G/B5G Networks Using Machine Learning: A Comparative Study. Sensors. 2023; 23(6):2967. https://doi.org/10.3390/s23062967

Chicago/Turabian Style

Chatzoglou, Efstratios, and Sotirios K. Goudos. 2023. "Beam-Selection for 5G/B5G Networks Using Machine Learning: A Comparative Study" Sensors 23, no. 6: 2967. https://doi.org/10.3390/s23062967

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop