Next Article in Journal
Analysis of Return-to-Zero Error after the First Load of Load Cell
Next Article in Special Issue
BézierCE: Low-Light Image Enhancement via Zero-Reference Bézier Curve Estimation
Previous Article in Journal
An Edge Transfer Learning Approach for Calibrating Soil Electrical Conductivity Sensors
Previous Article in Special Issue
Dataset Condensation via Expert Subspace Projection
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

DeepVision: Enhanced Drone Detection and Recognition in Visible Imagery through Deep Learning Networks

by
Hassan J. Al Dawasari
1,2,
Muhammad Bilal
1,2,
Muhammad Moinuddin
1,2,*,
Kamran Arshad
3,4,* and
Khaled Assaleh
3,4
1
Electrical and Computer Engineering Department, King Abdulaziz University, Jeddah 21589, Saudi Arabia
2
Centre of Excellence in Intelligent Engineering Systems, King Abdulaziz University, Jeddah 21589, Saudi Arabia
3
Department of Electrical and Computer Engineering, College of Engineering and Information Technology, Ajman University, Ajman 346, United Arab Emirates
4
Artificial Intelligence Research Centre, Ajman University, Ajman 346, United Arab Emirates
*
Authors to whom correspondence should be addressed.
Sensors 2023, 23(21), 8711; https://doi.org/10.3390/s23218711
Submission received: 28 August 2023 / Revised: 5 October 2023 / Accepted: 17 October 2023 / Published: 25 October 2023
(This article belongs to the Special Issue AI-Driven Sensing for Image Processing and Recognition)

Abstract

:
Drones are increasingly capturing the world’s attention, transcending mere hobbies to revolutionize areas such as engineering, disaster aid, logistics, and airport protection, among myriad other fascinating applications. However, there is growing concern about the risks that they pose to physical infrastructure, particularly at airports, due to potential misuse. In recent times, numerous incidents involving unauthorized drones at airports disrupting flights have been reported. To solve this issue, this article introduces an innovative deep learning method proposed to effectively distinguish between drones and birds. Evaluating the suggested approach with a carefully assembled image dataset demonstrates exceptional performance, surpassing established detection systems previously proposed in the literature. Since drones can appear extremely small compared to other aerial objects, we developed a robust image-tiling technique with overlaps, which showed improved performance in the presence of very small drones. Moreover, drones are frequently mistaken for birds due to their resemblances in appearance and movement patterns. Among the various models tested, including SqueezeNet, MobileNetV2, ResNet18, and ResNet50, the SqueezeNet model exhibited superior performance for medium area ratios, achieving higher average precision (AP) of 0.770. In addition, SqueezeNet’s superior AP scores, faster detection times, and more stable precision-recall dynamics make it more suitable for real-time, accurate drone detection than the other existing CNN methods. The proposed approach has the ability to not only detect the presence or absence of drones in a particular area but also to accurately identify and differentiate between drones and birds. The dataset utilized in this research was obtained from a real-world dataset made available by a group of universities and research institutions as part of the 2020 Drone vs. Bird Detection Challenge. We have also tested the performance of the proposed model on an unseen dataset, further validating its better performance.

1. Introduction

The use of drones for military, commercial, and security purposes is on the rise, as their manufacturing technologies continue to advance, leading to an increase in their numbers [1,2]. The efficacy of using various types of drones in airport security, facility protection, and integration with surveillance systems has garnered significant interest in the past few years [3]. Conversely, drones can pose a substantial risk in these security domains, underscoring the need to devise an effective method for detecting different drone varieties in such applications [4]. These technologies have the potential to safeguard airports and military installations against drone incursions or breaches, thereby enhancing their overall security [5]. Hence, it is imperative to discuss public safety and the potential risks posed by unmanned aerial vehicles (UAVs) by focusing on their detection, recognition, and identification. Detection involves observing a target, which may appear dubious and jeopardize the security of the surrounding environment. Recognition entails ascertaining the category to which the target belongs, while identification involves identifying the specific type of target category.
The main goal of this article is to detect and identify a specific drone variant by highlighting similarities in their physical attributes and behaviors to those of birds. Additionally, the research aims to distinguish this type of drone from birds. Various sensors have been utilized in literature for identical and detection purposes, including radar [6], LIDAR [7], and RF-based sensors [8]. Moreover, acoustic [9] and thermal [10] sensors have also been employed for drone detection and recognition. Nonetheless, the majority of such sensors are expensive and consume substantial amounts of energy [11]. Furthermore, the integration of drones with these sensors is restricted by their weight and size, and thermal sensors typically have lower resolution. Despite the absence of challenges related to incorporating detectors and unmanned aerial vehicles, employing optical imaging systems offers greater detail than infrared systems. However, challenges such as obscured regions, congested backgrounds, and inconsistent illumination in drone-captured images do exist. Consequently, the effectiveness of addressing these concerns depends on the particular approach used for detecting and distinguishing the aerial device. Identifying and discerning drones from birds is often challenging due to various factors, such as: (1) the similarities in physical appearance and behaviors between drones and birds, causing confusion, especially at long distances; (2) obstacles in detection due to environmental conditions, such as weather, crowds, and lighting variations; and (3) the difficulties in identifying small-sized drones at longer distances, resulting in challenges in recognition.
Deep learning networks have emerged as the most effective models for visual processing, particularly in tasks such as object detection and tracking [12]. Over the past decade, there has been significant interest in using these networks for object detection, as they offer greater accuracy and computational power than alternative approaches [13]. Deep neural networks, particularly convolutional neural networks (CNNs), have been widely regarded as among the most effective approaches for object recognition [14]. CNNs excel at feature extraction, making them a popular and extensively researched choice for object recognition [15]. Due to their superior ability to extract features, CNNs are preferred over traditional object recognition methods [16]. Therefore, a research study was carried out to assess the ability of CNNs to detect drones in visible images. Four different CNN models were utilized, namely ResNet-18, ResNet-50, MobileNet-V2, and SqueezeNet. ResNet-18 and ResNet-50 are particularly popular due to their ability to handle large datasets and complex features. MobileNet-V2 is a lighter-weight architecture designed for mobile devices, while SqueezeNet is a compressed architecture that provides good accuracy with fewer parameters. The purpose of using these specific models was to evaluate their individual strengths and weaknesses in identifying drones in images that are invisible to the human eye.
Detecting and recognizing drones are crucial for a variety of applications, including public safety. Visible imagery is an ideal sensor for this purpose due to its high resolution, low cost, and compatibility with different drones. However, this approach presents challenges, such as crowded backgrounds and confusion between drones and birds due to their similar sizes. To address these issues, a suitable method must be employed. Deep networks based on transfer learning are known for their ability to provide high accuracy and speed in analyzing input images [17]. Transfer learning involves a pre-trained model as a starting point and then fine-tuning it for a specific task. This approach enables the model to leverage knowledge gained from large datasets and use it for a new task, resulting in improved accuracy and faster training times. Transfer learning is particularly useful when working with limited training data, as it allows the model to learn from a broader dataset without requiring as much data for training. Therefore, transfer learning-based deep networks are an effective and efficient solution for image analysis tasks, including object detection, recognition, and classification. The objectives of this study were: (1) to develop an efficient and effective deep network for UAV detection; (2) to evaluate the performance of four popular CNN architectures, including ResNet-18, ResNet-50, MobileNet-V2, and SqueezeNet, for recognizing UAVs utilizing visible imagery; and (3) to achieve accurate and real-time identification of UAVs, with various practical applications, such as security, surveillance, and search and rescue operations. Our main contributions are as follows:
(1)
The proposed method is tested on a real-world dataset made available by a group of universities and research institutions as part of the 2020 Drone vs. Bird Detection Challenge;
(2)
The proposed method is able to detect and identify a specific drone variant by highlighting similarities in their physical attributes and behaviors with those of birds;
(3)
Given that drones can appear to be incredibly small in relation to other flying objects, we presented a robust image-tiling technique using overlaps that performed better in the presence of very tiny drones;
(4)
An innovative deep learning method is proposed to effectively distinguish between drones and birds, the main features of which are as follows:
(a)
The proposed method is based on the SqueezeNet model, which has less computational complexity than the best-performing ResNet18 model for the task of drone detection;
(b)
The proposed SqueezeNet-based method has shown better AP scores, faster detection times, and more stable precision-recall dynamics;
(c)
The proposed method is more suitable for real-time implementations because it has less computational architecture and a faster detection time;
(5)
The proposed method’s performance is further validated by testing its detection performance on an unseen dataset.

2. Materials and Methods

2.1. Dataset Description

The training dataset utilized for the Drone vs. Bird Detection Challenge comprises 77 distinct videos, each consisting of an average of 1384 frames, with an average of 1.12 annotated drones per frame. Among these videos, 11 were collected from the previous iteration of the challenge through MPEG4-coded static cameras by the Fraunhofer IOSB research institute and the SafeShore project. In 2020, the ALADDIN project contributed 45 additional videos in three different resolutions, namely 1920 × 1080 @25 fps, 720 × 576 @50 fps, and 1280 × 720 @30 fps. These videos were recorded at two different locations, each with distinct geographic characteristics, and they feature seven types of drones, including three with fixed wings and four rotary ones. Most of the videos are short, lasting around one minute, with the exception of two videos that are five minutes long. The remaining 21 training videos were supplied by Fraunhofer IOSB and were recorded at various locations throughout Germany, featuring five types of rotary drones. In total, the dataset consists of videos showcasing eight different types of drones.
In general, the training dataset demonstrates a high degree of variability in terms of difficulty. This variability results from the videos encompassing a diverse set of backgrounds, including sky or vegetation, and various weather conditions, such as sunny and cloudy weather. Additionally, the dataset also incorporates direct sun glare and variations in camera characteristics. Another aspect that adds to the complexity of the dataset is the varying distance of the drones from the camera, which can differ significantly within and across videos. As a result, there is a substantial variation in the sizes of drones. Drone sizes can range from as small as 15 pixels to more than 1,000,000 pixels. The majority of annotated drones in the dataset are relatively small, with sizes ranging from fewer than 162 pixels to 322 pixels. The small size of the drones in many of the videos makes the detection task even more challenging. The dataset includes both static and moving camera recordings. Each video is accompanied by a separate annotation file containing the frame number and the bounding box (expressed as [topx topy width height]) for the frames in which drones enter the scene. Some of the videos also feature birds, which are not individually annotated and can be a significant source of interference.
The evaluation dataset comprises an additional 14 videos that lack accompanying annotations. These videos exhibit characteristics akin to those in the training dataset, with some shared locations. However, not all the locations are identical. Within the evaluation dataset, three video sequences feature a moving camera. To prevent any single video from disproportionately influencing the evaluation results, longer sequences have been truncated to one minute. Access to the dataset requires signing a data utilization agreement (DUA), and the annotations can be found at https://github.com/wosdetc/challenge, accessed on 14 April 2021.

2.2. Split Dataset and Software

The videos were separated into two distinct groups for training and validation by the authors through a manual process. The primary objective of grouping was to maintain independence between the two groups to the greatest extent possible. The division was performed on the basis that videos recorded at identical locations must be included in the same group. The validation group was then chosen by selecting 16 videos from the split. The majority of the videos used for this study were captured at high frame rates, which led to a significant number of identical images. To address this issue, a subsampling technique was performed by selecting every 10th frame, resulting in a reduction of the dataset to only 10% of the original data. This smaller dataset was then used for training and validating the model.
Matlab 2022b software and deep learning modules were employed for the analysis. The experiments were conducted on a PC with an Intel Core i7-7500U processor running at 2.70–2.90 GHz and 12 GB of RAM. For the classification task, ResNet18, ResNet50, MobileNetV2, and SqueezeNet modules from the TensorFlow library, version 2.6.2, were utilized, all of which can be downloaded for free.

2.3. Image Preprocessing

This section details the dataset used for training and the methods used for image tiling and data augmentation. The primary dataset utilized to train and test the approach was the “2020 Drone vs. Bird Challenge”. Under the leadership of Alexis, the team removed one annotated sequence from the available 77, resulting in a total of 106,485 frames. Although the dataset includes both birds and drones, the annotations are limited to drones only. To further enhance the dataset, 26,500 synthetic images were added through the method outlined in the “Cut, Paste and Learn” paper [18]. To create these images, roughly 100 close-up images of birds and drones were sourced from Google Images, and object masks were manually generated. The objects were then blended onto outdoor backgrounds through various methods. In the synthetic dataset, annotations for bounding boxes were limited to drones only, unlike the Drone vs. Bird dataset, in which both drones and birds were annotated.
One of the key challenges in drone detection is the possibility of drones being exceptionally small, often appearing as merely 10–20 pixels compared to a full high-resolution image. Downsizing images before processing them through deep convolutional neural networks can be ineffective in detecting such small drones, as they may become too insignificant to differentiate from other aerial objects. As a solution, image tiling has proven to be a highly effective strategy [19], albeit at the cost of longer processing times. The image tiling approach involves dividing the input image into multiple patches or tiles. To enhance robustness and redundancy, overlaps may be included between tiles, which can be especially useful in scenarios in which drones appear on the borders of a tile.
During the training procedure, random cropping was utilized as a means of emulating image tiling. In 80% of cases, the cropping location was chosen in a way that ensured that at least one drone was included. In all other cases, the crop location was chosen randomly. This strategy helped to provide more instances of positive drone examples while also providing background examples. To improve the training process, the team led by Alexis utilized various data-augmentation techniques. These techniques included affine transformations, flips, blurring, noise, brightness adjustments, and several other augmentations sourced from the Albumentations package [20].

2.4. Pre-Trained Deep Networks

This article proposes a deep learning-based method for detecting and recognizing drones despite the challenges posed by crowded backgrounds, resemblances to birds, small sizes at longer distances, and lighting inconsistencies. The proposed approach involves four main steps, as shown in Figure 1. The first step involves preparing the input data for the proposed architecture. In the second step, a network training phase is implemented to detect and recognize two types of drones, as well as birds. The third step involves testing the trained model on a large variety of drone and bird datasets. Finally, the model’s performance is evaluated, and the detection and recognition process is conducted on the input test data.

2.4.1. Resnet18

ResNet-18 is a deep neural network architecture that was introduced by Microsoft Research in 2015 as a part of their research on deep learning. It is a variant of the ResNet (Residual Network) family of neural networks, which are known for their ability to train very deep neural networks utilizing residual connections [21]. ResNet-18 consists of 18 layers, including a convolutional layer, followed by four blocks of convolutional layers, and it ends with a fully connected layer. Each of these blocks contains two or three convolutional layers and a shortcut connection, which allows the network to learn residual functions. The use of residual connections in ResNet-18 allows for easier training of very deep neural networks by enabling the network to learn an identity mapping if needed, effectively bypassing a set of layers if it deems it necessary. In this way, the gradient can flow more easily through the network, and the network can learn to model more complex features. ResNet-18 has been widely used in various computer vision tasks [22], such as image classification, object detection, and segmentation. It has achieved state-of-the-art performance on many benchmark datasets, including ImageNet, CIFAR-10, and CIFAR-100.

2.4.2. Resnet50

ResNet is an advanced deep learning architecture consisting of fundamental residual blocks that serve as a connection across layers. This connection permits the training of hundreds or thousands of layers while simultaneously enhancing performance. ResNet was the winner of the 2015 Imagenet classification and is a dependable solution for many computer vision problems [23]. The number accompanying ResNet denotes the number of deep layers, with ResNet50 having 50 processing layers. The network’s size is 96 MB, with 25.6 million parameters. The ResNet50 architecture accepts input in 256 dimensions and features skip connections, which address the vanishing gradient descent problem. Each ResNet begins with convolution and max-pooling operations, followed by stacked convolutions. This layered convolution approach provides advantages in resolving the vanishing gradient problem [24].

2.4.3. MobileNetV2

The network based on depthwise separable convolutions was initially developed for mobile and embedded vision applications, as presented by Howard et al. [25]. Compared to other common models, deep convolutional neural networks, such as MobileNet, are more compact and faster. Each input is filtered by a single filter using depth-wise independent convolutions before being combined into a collection of output features through 1 × 1 convolutions. Depth-wise separable layers are similar to traditional convolutional layers, but they work much faster and with less modification. In depth-separable convolution, the standard filters and assemblies are separated into two layers: one for filtering and one for mixing. This process reduces the model’s size and computational power requirements. Batch normalization and ReLU nonlinearity are applied to all layers in this network, except the last fully connected layer, which feeds into a softmax layer for classification without nonlinearity. MobileNet’s convolutions include 28 layers, excluding the depthwise and pointwise convolutions. The network has a size of 5.2 MB and contains 1.3 million parameters.

2.4.4. SqueezeNet

SqueezeNet is a convolution network that executes better performance than AlexNet with 50× fewer parameters [26,27]. SqueezeNet consists of 15 layers with five different layers as two convolution layers, three max pooling layers, eight fire layers, one global average pooling layer, and one output layer softmax. SqueezeNet is a deep neural network architecture designed for efficient computation and small memory footprints. It was introduced by researchers at the University of California, Berkeley, in 2016. SqueezeNet uses a combination of techniques, such as filter factorization, 1 × 1 convolutions, and aggressive downsampling, to reduce the number of parameters while maintaining high accuracy. Compared to ResNet-18, SqueezeNet has a much smaller memory footprint and requires fewer computations, making it more suitable for deployment on devices with limited computing resources, such as drones. However, the trade-off is that SqueezeNet may not be as accurate as larger and more complex models, such as ResNet-18.

2.5. Evaluation Metrics

To assess the effectiveness of the proposed method, a range of evaluation metrics were used, including Intersection over Union (IoU), precision, and recall [28,29]. This evaluation strategy allowed for a more comprehensive understanding of the model’s performance to be gained. The IoU metric measures the extent of overlap between the predicted bounding box and the ground truth bounding box. In the study, the optimal threshold for classifying the input data was identified by plotting the true positive rate against the false positive rate for various thresholds and choosing the threshold that maximizes the area under the curve. The chosen threshold was 0.5, which was used to classify the input data. If the IoU value exceeds 0.5, the classification is considered a true positive (TP); otherwise, it is classified a false positive (FP). The confusion matrix was constructed using the numbers of TP, FP, TN (true negative), and FN (false negative) values.
The confusion matrix (Figure 2) is a fundamental tool for evaluating model accuracy and is typically of size n × n, where n is the number of classes [30]. In this study, the columns of the confusion matrix represent the true class of intended objects, including two types of drones and birds. Meanwhile, the rows indicate the predicted classes generated by the proposed deep learning model. The positive class pertains to drones, while the negative class relates to birds. Since this study involves two classes, the confusion matrix is generalized to a size of 2 × 2. Precision and recall are computed using the following equations:
I o U = T P F P + T P + F N × 100
P r = T P T P + F P × 100  
R e = T P T P + F N × 100

3. Results and Discussion

3.1. Ablation Study for Selection of SqueezeNet Hyperparameters

To determine the best hyperparameters empirically, an ablation study was performed using SqueezeNet as the baseline model. The findings of this study are reported in Table 1. The results are reported around the MiniBatch size, L2 regularization, and Momentum values of 2, 0.0005, and 0.9, respectively, yielding the best results.

3.2. Assessment of SqueezeNet and ResNet18 Performance

Table 2 presents a comparison of the performance of the deployed deep learning models, including SqueezeNet and ResNet18, across different overlap thresholds, area ratios, and training and detection times. A thorough investigation was conducted on the average precision (AP) for the range of overlap thresholds between 0.5 and 0.9. Additionally, the AP values were calculated for three different area ratios, namely small, large, and medium. The results unequivocally point to the SqueezeNet-based network, fortified by the use of data-augmentation techniques, as the most efficient in drone detection during the evaluation process, a finding also corroborated by Schmidhuber [31]. Specifically, SqueezeNet achieved an AP of 0.327 for overlap thresholds ranging from 0.5 to 0.9. This outcome is significant when contrasted with the ResNet18 model, which registered a comparatively lower AP value of 0.237 within the same range of overlap thresholds. Further granularity is provided when we look at area ratios. For medium-sized area ratios, the SqueezeNet model surpassed ResNet18 by securing an AP of 0.770 compared to ResNet18′s AP of 0.739. Although both models performed robustly for medium area ratios, SqueezeNet’s edge in this category could make it the preferred choice for specific drone-detection scenarios requiring that particular area ratio. Moreover, when it comes to detection time—an essential factor in real-time, critical applications—SqueezeNet was superior again, clocking in at a mere 0.061 s. This time is significantly shorter than the ResNet18 model’s time of 0.105 s, underscoring the computational efficiency of the former. Figure 3 deepens our understanding by illustrating the precision–recall relationship for these two architectures. As the recall rate increased, SqueezeNet maintained high precision levels, fluctuating between 100% and 88%. Meanwhile, ResNet18′s precision saw a larger degree of variance, ranging from 98% to 80%. These findings reinforce the importance of top-level network features in the effectiveness of drone detection. In summary, SqueezeNet’s superior AP scores, faster detection times, and more stable precision-recall dynamics indicate that it may be more suitable for real-time, accurate drone detection than ResNet18.

3.3. Performance Analysis of ResNet50 and MobileNetV2

Table 3 offers a comprehensive look at the comparative performance of two deep learning models, ResNet50 and MobileNetV2, focusing on various evaluation metrics, including different overlap thresholds, area ratios, and both training and detection times. At a glance, the results highlight the superiority of ResNet50 over MobileNetV2 when it comes to efficiently detecting drones. Diving deeper into the data, for the specified overlap thresholds spanning from 0.5 to 0.9, ResNet50 garnered a notable AP value of 0.254. In contrast, MobileNetV2 lagged behind, managing an AP score of only 0.166 within the same overlap threshold range. Such a significant disparity in scores underscores the refined efficiency of ResNet50 in these particular evaluation conditions. This pattern of dominance by ResNet50 continues when we factor in the area ratios. Particularly with medium area ratios, ResNet50 outpaced MobileNetV2. The former achieved an AP score of 0.610, while the latter settled for a score of 0.485. Such distinctions in performance can be pivotal, especially in applications in which accurate detection in medium-sized areas is of paramount importance. However, it is not all smooth sailing for ResNet50. In terms of sheer speed, MobileNetV2 showcased a detection time of a brisk 0.111 s, edging out ResNet50, which clocked in at 0.129 s. This outcome hints at the architectural efficiency of MobileNetV2, making it potentially more suitable for scenarios in which rapid detection is more crucial than utmost precision. Figure 4 further amplifies our understanding by offering a visual representation of the precision–recall dynamics for both models. As recall values rose, ResNet50 demonstrated precision ranging between 100% and 74%. MobileNetV2, while exhibiting a slightly broader fluctuation, maintained precision levels between 100% and 80%. Crucially, these findings serve as a testament to the significance of hyper-parameter tuning in the training phase. For instance, judiciously setting overlap thresholds and area ratios can critically enhance a model’s efficacy in drone detection. The sentiment aligns perfectly with the insights from the research by Redmon and Farhadi [32]. Their work spotlighted the instrumental role of selecting optimal hyperparameters, especially overlap thresholds and area ratios, in fine-tuning the performance of object detection models. This role remains true irrespective of the target detection object, but it is particularly vital in specialized tasks such as drone detection, in which precision can be mission-critical.

3.4. Proposed Deep Networks’ Learning Curves

Several investigations were conducted to enhance the learning curve of the super recognition model. These investigations included: (1) examining the properties of the selected premium features obtained by training a specific ML model [33]; (2) selecting optimal factors, including overlap thresholds and area ratios; (3) applying data augmentation techniques; and (4) training the deep network with early stopping to prevent overfitting. These efforts have culminated in the development of a high-performance deep network architecture. The enhanced SqueezeNet network surpassed other suggested networks, displaying reliable true positive and true negative values without any false negatives or false positives, making it effective in drone identification. As illustrated in Figure 5, it demonstrates the behavior of advanced models with iterations during training on the dataset. As the number of iterations increases, the learning rate gradually increases during training until a point is reached at which the learning curves indicate a high-quality model. Simultaneously, the loss of the models gradually decreases as the number of iterations increases. As depicted in Figure 5a–d, the performance of SqueezeNet, ResNet18, ResNet50, and MobileNetV2 improved after 5, 9000, 9000, and 6500 iterations, respectively. This study’s results match those of Elsayed et al. [34], who employed early stopping during the training of different deep networks to mitigate overfitting, leading to an anticipated enhancement of the expected performance of the models.

3.5. Performance Comparison

This section presents an assessment of the method’s effectiveness, in which experiments were conducted to demonstrate the influence of various hyper-parameters, including tile size and data-selection and augmentation approaches, which can yield significant effects. The evaluation employed the AP, which serves as the standard measure for object detection. The results displayed in Table 4 correspond to a test dataset consisting of images obtained through a varied methodology. Based on the AP scores presented in the table, SqueezeNet achieved the highest AP score of 0.770, indicating a strong ability to accurately classify the images in the test set. MobileNetV2 and ResNet18 achieved lower AP scores of 0.485 and 0.739, respectively, suggesting a relatively weaker performance compared to SqueezeNet. Interestingly, ResNet50 demonstrated an AP score of 0.610, which falls between the scores of MobileNetV2 and ResNet18. This outcome indicates that ResNet50 performed moderately well in classifying the images. Moreover, it is worth noting that the AP score reported by Coluccia et al. [35] in their study was significantly higher, at 52.2%. This finding implies that their approach, which is not explicitly mentioned in the provided information, outperformed all the methods examined in the current study. This research presents examples of successful detections through the suggested models illustrated in Figure 6. It should be noted that the dataset of the 2020 Drone vs. Bird Detection Challenge has a large number of images in which the drones are very small in size and almost indistinguishable from bird images, as can be seen in a few samples provided in Figure 6. All of the investigated models underwent evaluation, and their detection results are showcased alongside the ground truth annotations. Exploring the underlying factors that contribute to the varying performances of each algorithm on distinct sequences is an important area for further study. Upcoming iterations of the contest might assess other dimensions, such as computational complexity (and real-time capabilities), incorporation of supplementary datasets, and the capacity to adapt to entirely distinct backgrounds. These evaluations will help us to better comprehend the various compromises and interactions at play among the multitude of potential tactics and method combinations.

3.6. Proposed Model’s Validation on an Unseen Dataset

To validate the performance of our proposed method, it was investigated on an entirely different dataset. Figure 7 shows the detection results of the proposed model trained using Squeezenet as the backbone on sample test images taken from a different dataset [36]. The detection results prove that the trained model generalized well onto images from an entirely different dataset, as well as under different backgrounds. Only a few instances were not detected in certain images. These results may be reproduced using our developed open-source code provided at the following link: https://github.com/4mbilal/drone_detection (accessed on 3 August 2023).

4. Conclusions

The use of drones has become increasingly popular, leading to concerns about their potential use for nefarious purposes, particularly in sensitive locations such as airports. To address this issue, there has been significant interest in developing methods for detecting and recognizing drones. This issue is challenging due to the similarities between the behaviors and appearances of drones and birds in the sky, as well as the high speeds of drones, the presence of crowded backgrounds and hidden areas, lighting issues, and the difficulty of identifying small drones at long distances. To overcome these challenges, this paper proposes a novel deep learning-based approach for detecting and recognizing both drones and birds to prevent unauthorized drone activity. This study utilized a dataset sourced from a real-world dataset made accessible by a consortium of universities and research institutions as a component of the Drone vs. Bird Detection Challenge, held in 2020. Among the various models rigorously evaluated, including SqueezeNet, MobileNetV2, ResNet18, and ResNet50, the SqueezeNet architecture emerged as the standout performer. It exhibited exceptional performance, particularly for medium area ratios, boasting a remarkable average precision (AP) score of 0.770. While the initial results are promising, there remain several avenues for further exploration. In the future, there will be a significant emphasis on harnessing various deep learning networks to effectively differentiate drones from birds. This research, which encompasses diverse drone types, including multi-rotors, helicopters, fixed-wing, and VTOL aircraft, seeks to develop a real-time system capable of detecting, recognizing, and localizing these aerial objects, with integration into onboard systems as the ultimate objective.

Author Contributions

Conceptualization, M.B., M.M., K.A. (Kamran Arshad) and K.A. (Khaled Assaleh), methodology, M.B. and H.J.A.D.; software, H.J.A.D. and M.B.; validation, M.M., H.J.A.D. and K.A. (Kamran Arshad); formal analysis, H.J.A.D. and K.A. (Khaled Assaleh); investigation, H.J.A.D., K.A. (Kamran Arshad) and M.B.; resources, H.J.A.D. and M.B. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Publicly available datasets were analyzed in this study. This data can be found here: https://wosdetc2021.wordpress.com/drone-vs-bird-detection-challenge/.

Acknowledgments

This paper was supported by the Ajman University Internal Research Grant No. 2022-IRG-ENIT-22. The research findings in this paper are solely the responsibility of the author(s).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Shi, X.; Yang, C.; Xie, W.; Liang, C.; Shi, Z.; Chen, J. Anti-drone system with multiple surveillance technologies: Architecture, implementation, and challenges. IEEE Commun. Mag. 2018, 56, 68–74. [Google Scholar] [CrossRef]
  2. Anwar, M.; Kaleem, Z.; Jamalipour, A. Machine Learning Inspired Sound-Based Amateur Drone Detection for Public Safety Applications. IEEE Trans. Veh. Technol. 2019, 68, 2526–2534. [Google Scholar] [CrossRef]
  3. Gomez, C.; Purdie, H. UAV-based Photogrammetry and Geocomputing for Hazards and Disaster Risk Monitoring—A Review. Geoenviron. Disasters 2016, 3, 23. [Google Scholar] [CrossRef]
  4. Yaacoub, J.P.; Noura, H.; Salman, O.; Chehab, A. Security analysis of drones systems: Attacks, limitations, and recommendations. Internet Things 2020, 11, 100218. [Google Scholar] [CrossRef]
  5. Shvetsova, S.V.; Shvetsov, A.V. Ensuring safety and security in employing drones at airports. J. Transp. Secur. 2021, 14, 41–53. [Google Scholar] [CrossRef]
  6. Semkin, V.; Yin, M.; Hu, Y.; Mezzavilla, M.; Rangan, S. Drone Detection and Classification Based on Radar Cross Section Signatures. In Proceedings of the 2020 International Symposium on Antennas and Propagation (ISAP), Osaka, Japan, 25–28 January 2021; pp. 223–224. [Google Scholar]
  7. de Haag, M.U.; Bartone, C.G.; Braasch, M.S. Flight-test evaluation of small form-factor LiDAR and radar sensors for sUAS detect-and-avoid applications. In Proceedings of the 2016 IEEE/AIAA 35th Digital Avionics Systems Conference (DASC), Sacramento, CA, USA, 25–29 September 2016; pp. 1–11. [Google Scholar]
  8. Nguyen, P.; Ravindranatha, M.; Nguyen, A.; Han, R.; Vu, T. Investigating cost-effective RF-based detection of drones. In Proceedings of the 2nd Workshop on Micro Aerial Vehicle Networks, Systems, and Applications for Civilian Use, Singapore, 26–30 June 2016; pp. 17–22. [Google Scholar]
  9. Svanström, F.; Englund, C.; Alonso-Fernandez, F. Real-Time Drone Detection and Tracking With Visible, Thermal and Acoustic Sensors. In Proceedings of the 2020 25th International Conference on Pattern Recognition (ICPR), Milan, Italy, 10–15 January 2021; pp. 7265–7272. [Google Scholar]
  10. Andraši, P.; Radišić, T.; Muštra, M.; Ivošević, J. Night-time detection of UAVs using thermal infrared camera. Transp. Res. Procedia 2017, 28, 183–190. [Google Scholar] [CrossRef]
  11. Drozdowicz, J.; Wielgo, M.; Samczynski, P.; Kulpa, K.; Krzonkalla, J.; Mordzonek, M.; Bryl, M.; Jakielaszek, Z. 35 GHz FMCW drone detection system. In Proceedings of the 2016 17th International Radar Symposium (IRS), Krakow, Poland, 10–12 May 2016; pp. 1–4. [Google Scholar]
  12. Liu, H.; Qu, F.; Liu, Y.; Zhao, W.; Chen, Y. A drone detection with aircraft classification based on a camera array. IOP Conf. Ser. Mater. Sci. Eng. 2018, 322, 052005. [Google Scholar] [CrossRef]
  13. Ye, D.H.; Li, J.; Chen, Q.; Wachs, J.; Bouman, C. Deep learning for moving object detection and tracking from a single camera in unmanned aerial vehicles (UAVs). Electron. Imaging 2018, 2018, 4661–4666. [Google Scholar] [CrossRef]
  14. LeCun, Y.; Bengio, Y.; Hinton, G. Deep Learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
  15. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 2012, 25, 1097–1105. [Google Scholar] [CrossRef]
  16. Chandana, R.K.; Ramachandra, A.C. Real Time Object Detection System with YOLO and CNN Models: A Review. arXiv 2022, arXiv:2208.00773. [Google Scholar]
  17. Wurm, M.; Stark, T.; Zhu, X.X.; Weigand, M.; Taubenböck, H. Semantic segmentation of slums in satellite images using transfer learning on fully convolutional neural networks. ISPRS J. Photogramm. Remote Sens. 2019, 150, 59–69. [Google Scholar] [CrossRef]
  18. Dwibedi, D.; Misra, I.; Hebert, M. Cut, paste and learn: Surprisingly easy synthesis for instance detection. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 1301–1310. [Google Scholar]
  19. Unel, F.O.; Ozkalayci, B.; Cigla, C. The Power of Tiling for Small Object Detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Long Beach, CA, USA, 16–20 June 2019. [Google Scholar]
  20. Buslaev, A.; Iglovikov, V.I.; Khvedchenya, E.; Parinov, A.; Druzhinin, M.; Kalinin, A.A. Albumentations: Fast and flexible image augmentations. arXiv 2018, arXiv:1809.06839. [Google Scholar] [CrossRef]
  21. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 770–778. [Google Scholar]
  22. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards real-time object detection with region proposal networks. In Proceedings of the Advances in Neural Information Processing Systems 28 (NIPS 2015), Montreal, QC, Canada, 7–12 December 2015; Volume 28. [Google Scholar]
  23. Glorot, X.; Bengio, Y. Understanding the difficulty of training deep feedforward neural networks. In Proceedings of the Thirteenth International Conference on Artificial Intelligence and Statistics, JMLR Workshop and Conference Proceedings, Sardinia, Italy, 13–15 May 2010; pp. 249–256. [Google Scholar]
  24. Russakovsky, O.; Deng, J.; Su, H.; Krause, J.; Satheesh, S.; Ma, S.; Huang, Z.; Karpathy, A.; Khosla, A.; Bernstein, M.; et al. Imagenet large scale visual recognition challenge. Int. J. Comput. Vis. 2015, 115, 211–252. [Google Scholar] [CrossRef]
  25. Howard, A.G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, H. Mobilenets: Efficient convolutional neural networks for mobile vision applications. arXiv 2017, arXiv:1704.04861. [Google Scholar]
  26. Özyurt, F.; Sert, E.; Avcı, D. An expert system for brain tumor detection: Fuzzy C-means with super resolution and convolutional neural network with extreme learning machine. Med. Hypotheses 2020, 134, 2020. [Google Scholar] [CrossRef]
  27. Iandola, F.N.; Han, S.; Moskewicz, M.W.; Ashraf, K.; Dally, W.J.; Keutzer, K. SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and <0.5MB model size. arXiv 2016, arXiv:1602.07360. [Google Scholar]
  28. Akhloufi, M.A.; Arola, S.; Bonnet, A. Drones chasing drones: Reinforcement learning and deep search area proposal. Drones 2019, 3, 58. [Google Scholar] [CrossRef]
  29. Flach, P.A. The geometry of ROC space: Understanding machine learning metrics through ROC isometrics. In Proceedings of the Twentieth International Conference on Machine Learning (ICML-2003), Washington, DC, USA, 21–24 August 2003; pp. 194–201. [Google Scholar]
  30. Susmaga, R. Confusion matrix visualization. In Intelligent Information Processing and Web Mining; Springer: Berlin/Heidelberg, Germany, 2004; pp. 107–116. [Google Scholar]
  31. Schmidhuber, J. Deep learning in neural networks: An overview. Neural Netw. 2015, 61, 85–117. [Google Scholar] [CrossRef]
  32. Redmon, J.; Farhadi, A. YOLO9000: Better, faster, stronger. In Proceedings of the 2017 Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 7263–7271. [Google Scholar]
  33. Elmetwalli, A.H.; Mazrou, Y.S.; Tyler, A.N.; Hunter, P.D.; Elsherbiny, O.; Yaseen, Z.M.; Elsayed, S. Assessing the efficiency of remote sensing and machine learning algorithms to quantify wheat characteristics in the Nile Delta region of Egypt. Agriculture 2022, 12, 332. [Google Scholar] [CrossRef]
  34. Elsayed, S.; El-Hendawy, S.; Dewir, Y.H.; Schmidhalter, U.; Ibrahim, H.H.; Ibrahim, M.M.; Elsherbiny, O.; Farouk, M. Estimating the leaf water status and grain yield of wheat under different irrigation regimes using optimized two-and three-band hyperspectral indices and multivariate regression models. Water 2021, 13, 2666. [Google Scholar] [CrossRef]
  35. Coluccia, A.; Fascista, A.; Schumann, A.; Sommer, L.; Dimou, A.; Zarpalas, D.; Méndez, M.; De la Iglesia, D.; González, I.; Mercier, J.P.; et al. Drone vs. Bird Detection: Deep Learning Algorithms and Results from a Grand Challenge. Sensors 2021, 21, 2824. [Google Scholar] [CrossRef] [PubMed]
  36. Drone Dataset (UAV). Available online: https://www.kaggle.com/datasets/dasmehdixtr/drone-dataset-uav (accessed on 20 August 2023).
Figure 1. The proposed detection and recognition diagram using different deep networks.
Figure 1. The proposed detection and recognition diagram using different deep networks.
Sensors 23 08711 g001
Figure 2. Confusion matrix of the proposed framework.
Figure 2. Confusion matrix of the proposed framework.
Sensors 23 08711 g002
Figure 3. The precision–recall relationship of proposed deep networks: (a) SqueezeNet; and (b) ResNet18.
Figure 3. The precision–recall relationship of proposed deep networks: (a) SqueezeNet; and (b) ResNet18.
Sensors 23 08711 g003
Figure 4. The precision–recall correlation of the proposed deep networks: (a) ResNet50; and (b) MobileNetV2.
Figure 4. The precision–recall correlation of the proposed deep networks: (a) ResNet50; and (b) MobileNetV2.
Sensors 23 08711 g004
Figure 5. The associations among learning rate, total loss, and iteration for the suggested networks: (a) SqueezeNet; (b) ResNet18; (c) ResNet50; and (d) MobileNetV2.
Figure 5. The associations among learning rate, total loss, and iteration for the suggested networks: (a) SqueezeNet; (b) ResNet18; (c) ResNet50; and (d) MobileNetV2.
Sensors 23 08711 g005aSensors 23 08711 g005b
Figure 6. Visual samples displaying qualitative detection outcomes alongside their corresponding ground truth annotations (yellow) with different proposed models: (a) SqueezeNet; (b) ResNet18; (c) ResNet50; and (d) MobileNetV2.
Figure 6. Visual samples displaying qualitative detection outcomes alongside their corresponding ground truth annotations (yellow) with different proposed models: (a) SqueezeNet; (b) ResNet18; (c) ResNet50; and (d) MobileNetV2.
Sensors 23 08711 g006
Figure 7. Visual samples displaying qualitative detection of the proposed model on an unseen dataset [36].
Figure 7. Visual samples displaying qualitative detection of the proposed model on an unseen dataset [36].
Sensors 23 08711 g007aSensors 23 08711 g007b
Table 1. A summary of the ablation study for selection of the SqueezeNet Hyperparameters.
Table 1. A summary of the ablation study for selection of the SqueezeNet Hyperparameters.
Hyperparameter A P 0.5 : 0.9 A P 0.5 A P 0.75 A P S A P M A P L
MiniBatch Size10.235810.125130.125130.509520.283930.013993
20.3270.2190.2190.6310.7700.216
40.263370.148320.148320.554190.6390.083013
80.229580.0981460.0981460.520130.487460.03198
160.147260.0375110.0375110.387040.210770.001357
L2 Regularization0.000250.0882290.0143260.0143260.255570.0015690
0.00050.3270.2190.2190.6310.7700.216
0.0010.254230.132650.132650.541140.696980.087404
Momentum0.950.1620.0950.0950.510.670.182
0.90.3270.2190.2190.6310.7700.216
0.850.178740.054530.054530.444490.317320.010222
Table 2. A comparison between the two architectures across different overlap thresholds, area ratios, and operation times.
Table 2. A comparison between the two architectures across different overlap thresholds, area ratios, and operation times.
BackboneTiTtTs A P 0.5 : 0.9 A P 0.5 A P 0.75 A P S A P M A P L
SqueezeNet0.319332.9350.0610.3270.2190.2190.6310.7700.216
ResNet180.455117.2090.1050.2370.0820.0820.5570.7390.264
where Ti is average time per single iteration (sec), Tt is time for training the network (min), Ts is time for single detection (sec), A P 0.5 : 0.9 represents the average precision calculated using an overlap threshold that ranges from 0.5 to 0.9, and A P S ,   A P L ,   and   A P M denote the average precision values for small, large, and medium area ratios, respectively.
Table 3. The performance of the two architectures across different overlap thresholds, area ratios, and operation time.
Table 3. The performance of the two architectures across different overlap thresholds, area ratios, and operation time.
BackboneTiTtTs A P 0.5 : 0.9 A P 0.5 A P 0.75 A P S A P M A P L
ResNet501.078273.7290.1290.2540.1040.1040.5790.6100.214
MobileNetV20.861219.6660.1110.1660.0360.0360.4420.4850.079
Table 4. Differences in results between this study and prior research.
Table 4. Differences in results between this study and prior research.
MethodSqueezeNetMobileNetV2ResNet18ResNet50(Coluccia et al., 2021) [35]
AP (%)0.7700.4850.7390.61052.2
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Al Dawasari, H.J.; Bilal, M.; Moinuddin, M.; Arshad, K.; Assaleh, K. DeepVision: Enhanced Drone Detection and Recognition in Visible Imagery through Deep Learning Networks. Sensors 2023, 23, 8711. https://doi.org/10.3390/s23218711

AMA Style

Al Dawasari HJ, Bilal M, Moinuddin M, Arshad K, Assaleh K. DeepVision: Enhanced Drone Detection and Recognition in Visible Imagery through Deep Learning Networks. Sensors. 2023; 23(21):8711. https://doi.org/10.3390/s23218711

Chicago/Turabian Style

Al Dawasari, Hassan J., Muhammad Bilal, Muhammad Moinuddin, Kamran Arshad, and Khaled Assaleh. 2023. "DeepVision: Enhanced Drone Detection and Recognition in Visible Imagery through Deep Learning Networks" Sensors 23, no. 21: 8711. https://doi.org/10.3390/s23218711

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop