Abstract
A mask identification and social distance monitoring system using Unmanned Aerial Vehicles (UAV) in the outdoors has been proposed for a health establishment. The above approach performed surveillance of the surrounding area using cameras installed in UAVs and internet of things technologies, and the captured images seem useful for tracking the entire environment. However, innate images from unmanned aerial vehicles show an adaptable visual effect in an uncontrolled environment, making face-mask detection and recognition harder. The UAV picture first had to be converted to grayscale, then its contrast was amplified. Image contrast was improved using Optimum Wavelet-Based Masking and the Enhanced Cuckoo Methodology (ECM). According to the contrast-enhanced image, Gabor-Transform (GT) and Stroke Width Transform (SWT) methods are used to derive attributes that help categorise mask-wearers and non-mask-wearers. Using the retrieved attributes, a Weighted Naive Bayes Classification (WNBC) detected masks in the images. Additionally, a deep neural network-based, the faster Region-Based Convolutional Neural Networks (R-CNN) algorithm combined with Adaptive Galactic Swarm Optimization (AGSO) is being used to identify appropriate and incorrect face mask wear in images, as well as to monitor social distancing among individuals in crowded areas. When the system recognises unmasked individuals, it sends their information to the doctor and the nearby police station. One unmanned aerial vehicle’s automated system alert people via speakers, ensuring social spacing. The problem involves a large percentage of appropriate and incorrect face mask wear using data from GitHub and Kaggle, including a training repository of 16,000 images and a testing data set of 12,751 images. To enhance the performance of the model’s learning, the methodology of 10-fold cross-validation will be used. Precision, recall, F1-score, and speed are then measured to determine the efficacy of the suggested approach.
Similar content being viewed by others
Introduction
The global COVID-19 pandemic has to be recognized as a serious threat. Considering the ongoing rise in occurrences, it is impractical for a governing body to keep track of them alone. The decrease in incidence of such an illness can only be possible with effective individual teamwork. The efficiency of minimizing transmission of viruses has been demonstrated extensively through the implementation of strategies such as distance from one another, hand sanitization, and the use of facial masks. Nonetheless, it is obvious that not all people follow these rules and regulations. Significant development is currently being made in the areas of Deep Learning (DL), Machine Intelligence (MI), Internet of Things (IoT)1, and Unmanned Aerial Vehicle (UAV). These developments possess the potential to accurately represent different global aspects, especially how populations are affected by physical barriers and the probability of individuals wearing face masks, among other things2,3,4,5. As a result, it is essential to develop an approach for dealing with issues that mask noncompliance and inattention to social distancing guidelines.
The IoT6technological advances sector has made substantial advancements, improving one’s capacity to carry out remote monitoring. The interconnectedness of computer-assisted elements across an internet7is an important aspect of the Internet of Things. To facilitate the transfer of data, each component will be allocated a unique role. Specifically, the method involves developing a face mask detection system using IoT technologies8. The proposed approach employs a dataset of images captured by an unmanned aerial vehicle (UAV) in congested regions to detect and distinguish between individuals wearing face masks or not, as well as measure the social distance between individuals. While the system recognizes a face without a mask, it promptly transmits details containing the individual’s image to each of the healthcare providers9and the police station nearest them. The process of identifying face masks in UAV camera images involves two tasks: detecting the mask and identifying whether it is appropriate or incorrect But prior to that, the captured UAV image must be beforehand processed. During initial processing, an image has become enlarged. The pre-processed visuals have been subsequently filtrated using a dynamic, unsharp masking approach. This approach employs an enhanced cuckoo search optimization methodology to select approximate coefficients. The application of an enhanced cuckoo search algorithm culminated in the design of an effective wavelet filter10, which is capable of constantly increasing contrast in images.
The image with higher contrast will be separated using the marker-controlled watershed segmentation method. The outcomes of segmentation are used to extract features via Gabor’s transform (GT). The features extracted are subsequently put into weighted Naive Bias Classification (WNBC) methodology to identify visuals with and without wearing face masks among the available UAV captured images. subsequently determining these two images, the one with the face mask is input into the faster RCNN, which combines with the OpenCV library. The faster R-CNN suggests using bounding boxes to determine the region of interest11 in order to distinguish between acceptable and erroneous facemask wear in UAV images. Furthermore, throughout the entire procedure of detecting appropriate and incorrect face mask wear in UAV images, whenever the data set comprises intricate, deformed images of individuals, the faster RCNN paired with the open OpenCV library operates less accurately. In order to prevent such circumstances, an AGSO is used along with faster RCNN to choose the most suitable weight parameter for faster RCNN. This enhances the overall accuracy of the entire procedure and emphasizes the novel component of the proposed methodology. In addition, the proposed approach is focused on monitoring social distancing among individuals using object recognition and traceability methods. the distance between the individuals is recognized in real time using bounding boxes, Raspberry Pi 4, OpenCV webcam, and UAV. The article is presented as follows: Sect. 2 explores the related work for the proposed approach. Section 3 describes the proposed methodology, while Sect. 4 examines the results and discussion. Finally, a conclusion has been defined in Sect. 5.
Related work
Individuals have been devising numerous innovative ways to protect themselves from the COVID-19 disease. Scientists in fields such as the internet-of-things, machine learning12, computer-vision, blockchain, and others will be working on numerous methods to protect and cure people from the spread of this viral infection. In this paper, research was carried out using IoT13, a deep learning-based faster R-CNN method14, to identify face-masked individuals and also measure the social distancing among individuals. The whole section will explain related research in the area. As stated in15 scientific article, the objective is to make a face-mask, and also a social distancing detection framework as an entrenched imaging system. Throughout our sense, pre-trained models like MobileNet, ResNet Classifier, and VGG have been often chosen. Individuals who violated social detachment and were not donning face-mask have been nabbed. This article also includes a comparison of various face-mask identification and classification techniques.
To protect individuals from COVID-1916, employs a creative smart method based on the deep convolutional approach17. The suggested system will instantly discern whether or not individuals strictly adhere to safety procedures. A comparative evaluation of numerous deep learning18strategies to monitor illness from healthcare image analysis19has been completed throughout20. Further analysis21 could very well suggest an internet of things device for COVID-19 safeguards related to thermal identification, face-mask recognition, and social-distancing. The Arduino Uno is being used for infra-red sensing devices, whilst the Raspberry Pi has been used for mask-detection as well as social-distancing via vision-based methodologies.
22proposes a new face mask detection method based on image pre-processing, face-detection, and imaging super-resolution. When it comes to identifying balaclava wearers, the framework has proven to be highly precise. In23, a deep neural network-based framework for monitoring individuals and social distancing among individual people has been proffered, even in low-light conditions. In aspects of speed as well as accuracy, a methodology has been proven to be preferable to several previous methodologies. In24, A computer-vision-based methodology has been used to probe mask detection and social distancing in real-time. To supervise the various operations, the concept has been built on the Raspberry Pi. It has been discovered that it can save time but also decrease the propagation of COVID-19.
In25, for COVID-19 detection, an IoT-based deep learning has already been provided. The offered outline would have been used to identify pandemics besides utilising it for X-rays of the chest. This has been demonstrated to be exceedingly exact, providing healthcare workers with the timely identification of COVID-19. To identify the individuals’ wearing masks, describes a method predicated on Open-Source Computer Vision. The method has been proven to be effective, particularly in industrial uses. A thorough study of the predominant facets of COVID-19 has been proffered in26, and also the impacts of COVID-19 on Artificial Intelligence (AI), Unmanned Aerial Vehicles (UAVs), the Internet of Things, 5G, and other technologies will be addressed.
In27, the involvement of IoT as well as the relevant sensor systems for coronavirus monitoring and mitigation is discussed. This same research delves deeply into the e-health services premised on sensing devices for handling COVID-19, as well as successive networks again for the post-pandemic age28. discusses an evaluation of a technique that could be used to identify COVID-19. The review also describes the difficulties that will be encountered in incorporating such techniques in coming years. Deep learning associated with X-ray, In-Vitro Diagnostics (IVDs), and IoT-based devices were amongst the advancements considered towards following COVID-19 patients.
The paper29describes the use of deep learning based methodologies to solve few main problems like trying to detect people with or without masks, as well as predicting whether or not social distancing would be noticed as just an image segmentation issue. The exploratory results affirm HRNetV2’s ability to generate forecasts for assessing social distancing scores. A paper will be particularly useful when comparing the computational achievement of various edge inferencing methodologies, as well as the overall accuracy of such methodologies for such issue domains. The bench-marking methods used throughout the paper focus on secondary information and exploratory runs utilising established bench-marking techniques. Significant research will have been conducted in the fields of object recognition30 and conceptual segmentation using deep learning techniques.
Proposed methodology
Face recognition is becoming increasingly important in content-based computer vision applications such as pattern classification31, monitoring, machine vision, fraud prevention, behavioural science, and neural nets32as a result of the global pandemic. During the investigation and competition, facial mask recognition likely garnered significant attention33. Mobile phone face-locking, signing with one’s face, and other applications have utilised face identification technology34. However, the identification of small and oblique face masks has been necessary in specific situations, such as areas with high crowd movement analysis, among others. Consequently, the pixel intensities in the image containing profiles and small faces are also incorporated. Convolutional neural networks (CNNs) are currently demonstrating enhanced performance in computer vision tasks, including object recognition, image labelling, and others35.
Consequently, scientists have employed these methodologies to implement several face recognition techniques36,37, such as R-CNN, fast R-CNN, and faster R-CNN. It enhances the generation of more timely and accurate discoveries. The proposed method illustrated in the Fig. 2, utilises the OpenCV38library along with the Raspberry Pi to develop an Internet of Things (IoT) system for identifying face masks. Additionally, it incorporates an UAV system for tracking social distancing39. A proposed system could effectively identify or verify an individual wearing mask or not by using a recorded video frame. The image40undergoes pre-processing, specifically employing a dynamic un-sharp masking technique. This technique involves selecting approximate coefficients using an optimisation methodology41. The Discrete Wavelet Transform (DWT) provided by Eq. 1 has been used to divide the original image into separate frequency bands that contain both accurate and approximate details. The levels of division appear to be categorised as Low-Level (LL), Low-High (LH), High-Low (HL), and High-High (HH). This encompasses three levels of comprehensive information, namely LH, HL, and HH, along with each level of approximation, LL. The low-level decomposition (LL) contains the highest amount of image data. The LH component contains the horizontal data, the HL component contains the vertical data, and the HH component contains the diagonal details. Following the initial stage of decomposition, the resulting elements are LL1, LH1, HL1, and HH1, respectively. LL1 has been selected for another round of decomposition in the exact same manner as depicted in the current image. The coefficient of approximation, LL2, will also be obtained in the next round. Similarly, an image is divided into up to four levels using the two-dimensional decomposition method. The obtained approximation coefficients, namely LL1, LL2, LL3, and LL4, as depicted in Fig. 1, are subsequently upscaled using advance cuckoo search algorithms and utilised in the construction of the filtration system to enhance the accuracy of the main image. The image would be decomposed into a maximum of 4 tiers due to the significant loss of detailed information after the 5th tier.
The coefficients approximated with the Inverse Discrete Wavelet Transformation (IDWT) as described in Eq. 2 have been calculated using Eq. 1.
$$\:W\phi\:(Jo, m, n)=\:\frac{1}{\sqrt{MN}}\sum\:_{x=0}^{M-1}\sum\:_{y=0}^{N-1}f\left(X,Y\right)\phi\:Jo,m,n\,(X, Y)$$(1)$$\:{f}_{(x,\:y)}^{A}=\:\frac{1}{\sqrt{MN}}{\sum\:}_{m}{\sum\:}_{n}w\varphi\:(Jo,m,n)\varphi\:Jo,m,n\,(X, Y)$$(2)
The approximation coefficients are \(\:W\varphi\:(Jo,m, n)\); J₀ is the arbitrarily defined preliminary scale valuation; m, n, are the wavelet realm discrete variables. f(x, y) is the temporal domains representation containing discrete variables as input x, y of size M*N, J₀=0,φ J₀,m, n(X, Y) is the scale-function, and \(\:{f}_{(x,y)}^{A}\) are the recreated approximation coefficients. For the filter formation a wavelet approximate image has been used rather than original image.
The utilisation of an enhanced cuckoo search algorithm has resulted in the development of an efficient wavelet filter that is capable of dynamically enhancing image contrast. The image with improved contrast would be divided using the marker-controlled watershed technique. The segmentation results are used to extract features using Gabor’s Transform (GT). These extracted features are then inputted into the weighted Naive Bias Classification (WNBC) to determine the presence or absence of a facemask. The Region of Interest (ROI) is acknowledged if it is present. Subsequently, ROI pooling has been employed to collect and resize image features for the purpose of creating a novel feature extraction42. Consequently, it was determined to utilise classification43and regression on each region of the new feature map in order to predict component values for the bounding box once more. R-CNN exhibit a lower object detection capability44compared to faster RCNN. Moreover, alternative connectivity is employed to identify a specific area of recommendation in specific cases of faster R-CNN. Due to the time-consuming nature of R-CNN and its additional computing requirements45, a more advanced iteration called faster R-CNN was created. In order to compensate for aspect ratio variations, faster R-CNN employs OpenCV bounding boxes to figure out the region of interest in UAV images for recognizing proper and erroneously face mask wear.
Although OpenCV was originally developed to provide a common framework for machine learning methods46and improve the practical application of machine perception and recognition, the OpenCV library has demonstrated remarkable effectiveness in identifying and recognising proper and not proper wear of the masks. The library encompasses over 2500 optimisation techniques47, comprising both conventional and state-of-the-art deep learning as well as computer-vision methodologies. The OpenCV libraries have employed optimisation methodologies to perform various tasks such as face image recognition48, interpretation and classification of living organism movement in videos, artefact recognition, webcam motion tracking, 3D approach extraction, creation of point clouds using stereo camera systems, image stitching to produce a composite image, identification of related images49 from a database, removal of bloodshot eyes from images, and assessment of physical movements such as eye movements, palpation, and gait. However, if the data set comprises complex, perverted framed visuals of people, the faster RCNN teamed with the open OpenCV library approach utilized for recognizing appropriate and erroneously face-mask wear within UAV images appears to operate less accurately. In order to prevent these kinds of scenarios, an AGSO is utilized with a faster RCNN to figure out the optimal weight parameter to be used for the faster RCNN. AGSO’s parameters have been continually adjusting employing a fuzzier technique. This AGSO method has been motivated by the shifting motion of stars and galaxies under the impact of the force of gravity. The AGSO works in two phases: exploration and extraction. Throughout the exploration stage, the subpopulation of particles searches within vector space for a best-case scenario. during the exploitation stage, the most appropriate solution for every subpopulation moves closer to the globally most effective. Also, such galaxies are treated as points of mass on an adequate scale. Once again, these achieved galaxies come together in large numbers in order to get the galaxies extremely grouped. As a result, the overall accuracy of identifying proper and erroneous face mask wear in UAV images improves.
In this work, it was found that a UAV primarily used for aerial imagery has the potential to replace certain parts in order to facilitate amphibious operations. Additionally, a Raspberry Pi kit can be used for real-time picture identification. It has potential applications in real-world scenarios such as data collection studies, aerial photography, safety regulations, and tracking50community distances in crowded areas51. The UAV could effortlessly discern the spatial arrangement among individuals and objects, including various entities such as swans, animals, plants, vegetation, and houses. This serves as an intimidation to government officials and individuals in positions of authority.
Unmanned Aerial Vehicles (UAVs), particularly when combined with the integration of the “Internet of Things” (IoT), are regarded as dependable aerial platforms for efficiently gathering and monitoring data at an affordable rate. It is employed to look at and manage physical surroundings and transmit data to information systems. The methods and procedures of the proposed methodology have been laid out in Table 1.
Results and discussion
In order to make sure of the fact that the outcomes had only been pertinent to the identification model, every experiment of the proposed approach has been carried out on the MATLAB platform alongside TensorFlow and the PyTorch deep learning framework using an Intel(R) Core (TM) i7-8565U CPU @ 1.80 –1.99 GHz, RAM 8GB, Ubuntu 18.04 operating system. A variety of hyperparameters have been indicated for experimentation: dimensions = 20, population size = 20, iteration 1 = 300, iteration 2 = 300, learning rate = 0.5204, gradient = 0.1131, weight decay = 0.05131, and epoch number = 100, via the best training performance of 0.0978 at epoch number 81. Given an epoch number of 81, the proposed methodology provides the best training performance, with a mean square error of 0.097828. When the number of epochs exceeds 81, the run time increases while the mean square error does not decrease significantly. As a result, every experiment needed an epoch of 81. Furthermore, the mean square error of proposed approaches increases as the number of epochs exceeds 81. Throughout this section, multiple factors, including validation loss, validation accuracy, precision, recall, and F1-score, have been evaluated. A specific approach is being compared to other well-established methodologies applied to the same set of images under different circumstances. The work will present the final outcomes regarding the impact of wearing face masks, not wearing face masks, practicing social distancing, and the prevalence of individuals using authorised streaming video services.
Dataset description
The proposed methodology utilizes a substantial amount of appropriate and incorrect face mask wear data sets collected through GitHub and Kaggle, which include the training repository of 16,000 images and a testing data set of 12,751 images. Each image has a variety of dimensions and backgrounds. For more effective recognition, images are captured in a variety of locations and lighting circumstances. To enhance the efficacy of the model’s training, the 10-fold cross-validation method is going to be used. The standard image data set illustrating appropriate and incorrect face mask wear has been thoroughly divided into ten separate groups using the ten-fold technique. The remaining subgroup is included in the test data, while the set of training data for every iteration of the models is made up of nine subgroups. To reduce bias, a substantial amount of the available data set is utilized to train the models over ten iterations. Furthermore, the model weights of the convolutional layers are constantly altered during every iteration, thus increasing the learning procedure’s efficiency. When the total amount of visuals is low, model accuracy suffers, along with the risk of overfitting. This suggested methodology employs data augmentation techniques to expand the dataset’s size in order to prevent these issues. Data augmentation techniques have significance for datasets with few images because they improve not only the size of the dataset but also the variety of input images, causing the model that was created able to adapt to complicated backgrounds and enhancing its robustness. Color change and geometric transformation techniques are frequently employed in data augmentation. Color transformation approaches include contrast, hue, and saturation transformation, while geometric transformation methods consist of random scaling, flipping, and cropping. Table 2 displays a range of techniques for face covering in the dataset, along with the amount of data used for training and testing. Figure 3 exhibits diverse categories within a test dataset. Figure 4 depicts the model’s training performance at its best, while Fig. 5 showcases the variability of validation accuracy for different training sets.
Validation loss
The training loss is used to measure the inaccuracies of the training dataset mentioned earlier, and it is evaluated at each epoch. The validation inaccuracy primarily stems from the transfer of the validation dataset during the training process, specifically assessed after each epoch. An epoch is a term used to describe a series of operations or iterations carried out on the complete training dataset. Equation (1) can be used to compute the validation loss using the training loss.
The loss in training and validation are evaluated to some extent based on the mean square error. Table 3 presents the training and validation loss for a multitude of training images. It is evident that the validation error remains constant or increases while the training error decreases from 12,000 to 2000 images. Indeed, the term “overfitting” is used to describe such a circumstance. Everything is in order after the completion of the 12,000-test set, as both are progressing along the same trend. Furthermore, Fig. 6 demonstrates the relationship between the validation loss and the number of training samples. The validation loss initially decreases as the number of training data images increases, but it significantly increases once it reaches 14,000.
Accuracy in validation
The accuracy of validation is a crucial metric for forecasting algorithm effectiveness. A model has been validated by assessing both the loss and consistency metrics. Table 5 displays, a comparison of Validation loss and Validation Accuracy of Proposed Methodology. The validation accuracy, as indicated in the last column, has consistently been below 1. This value is expressed as a percentage by multiplying the given evaluation by 100. The variability of validation accuracy persists, despite the reduction of the validation deficit through the augmentation of training examples from 4000 to 14,000. Consequently, the aforementioned paragraph has acquired the label of “overfitting.” In addition, as the quantity of pictures surpasses 16,000, the level of accuracy gradually diminishes. It indicates that the framework has been thoroughly mastered. Figure 7 illustrates the relationship between the accuracy of validation and the increasing number of training data images.
Performances metrices
The suggested methodology utilises True Positives (TP), False Positives (FP), True Negatives (TN), and False Negatives (FN) to compute performance metrics such as precision and recall52. Precision is the quotient of accurately predicted favourable experimental data divided by the total number of accurately predicted positive observations, while recall is the proportion of accurately predicted positive observational data to the total number of samples. The F1-score was computed by taking the harmonic mean of the recall and precision.
.
Table 5 illustrates the precision, recall, and F1-score for a specific test set. The dataset labelled “With Mask” includes the data sets numbered 1, 6, 15, and 16 in Table 2. The dataset labelled “Without Mask” includes all the remaining data sets. Figure 8 provides a visual representation of the comparison between precision, recall, and F1-score for testing datasets with and without wearing masks. The new methodology demonstrates excellent precision, recall, and F1-score for both types of datasets.
Evaluation of the suggested methodological approach to the available method
The proposed approach based on Faster R-CNN, has demonstrated superior performance when compared to other methods applied to the same dataset. Table 6presents a comparative analysis of speed, measured in frames per second. The data collection process will be divided into four datasets, namely Dataset 1, Dataset 2, Dataset 3, and Dataset 4 as per Fig. 9 . These datasets correspond to a college campus, a traffic area, a street, and a shopping mall, respectively. Studies have demonstrated that Faster R-CNN outperforms other techniques, including GravNet and DGCNN approaches53. Figure 8 illustrates a comparative assessment of the speed of concepts in frames per second (fps) between Faster R-CNN and other methods.
Outcomes for face-masking, social distancing, and counting in real time
The majority of UAVs are equipped with artificial intelligence and deep learning algorithms, which enhance their ability to perform tasks and increase their accuracy. Advanced unmanned aerial vehicles (UAVs) are able to obtain aerial imagery within real-time constraints. The utilisation of UAVs can greatly decrease the duration of real-time computing processes. The utilisation of libraries such as OpenCV, PyTorch, Tensor Flow, and Keras proved indispensable in effectively resolving live video reliability concerns and providing assistance in addressing object recognition challenges. Figure 10illustrates the instantaneous representation of the mask and unmask numbers. The UAV-captured image would undergo a conversion to grayscale, followed by an enhancement of its sharpness. Image contrast in the initial phase has been attained through the utilisation of Optimal Wavelet-Based Masking (OWBM) and the Enhanced Cuckoo Search Algorithm (ECSA). The distinguishing characteristics for differentiating individuals who wear masks from those who do not wear masks have been extracted from enhanced image frames in the second stage, employing Gabor transform and stroke width transform algorithms. In the third stage, a weighted naive Bayes classifier (WNBC) was utilised to accurately identify the presence of a face mask within a scene image based on the derived characteristics. The fourth stage utilises the faster R-CNN54technique, which is a deep neural network-based convolutional neural network. It employ AGSO to identify and rectify instances of incorrect mask donning in UAV images. Similarly, the process of measuring distance took place by employing detection and tracking technologies55, where each individual can be recognised in real-time using bounding box coordinates. Figure 10 illustrates the mask-wearing status, with green indicating the absence of a mask and red indicating proper mask usage. Figure 11 illustrates the process of determining social distance in a Kaggle dataset. The presence of red artefacts indicates that these artefacts have a relatively short distance and are not safe to approach.
Technical challenges faced during implementation
The main challenge faced during implementation is the dataset of images used in this proposed method, which exhibits limited variation within each class and low pixel density. Optimising the weights to achieve the most effective solutions and creating a highly functional and efficient Convolutional Neural Network (CNN) with minimal or maximal impact on the variables involved in the error function is a laborious task due to the complex landscape of ups and downs. There is also a technological challenge in the implementation process that needs to be solved by doing more research to get better frame data sets using OWBM and the Enhanced Cuckoo Search Algorithm (ECSA).
Conclusion
The whole world has undeniably experienced significant repercussions due to the outbreak of the coronavirus disease in 2019. It is advisable for individuals to wear a mask in public spaces as a substantial measure of protection. In order to impede the transmission of the virus, numerous regions implemented a compulsory policy requiring individuals to wear masks in densely populated areas. Multiple researchers have conducted investigations on image-based automated facial mask identification. This work involves the development of a sophisticated Unmanned Aerial Vehicle (UAV) that utilises advanced deep learning techniques and an improved version of the cuckoo optimisation algorithm. The primary objective of this UAV is to detect masks and track social distancing. Without an Unmanned Aerial Vehicle (UAV), it becomes more challenging to differentiate between a group of individuals who are genuinely not wearing masks and those who are not adhering to social distancing measures. A UAV, or unmanned aerial vehicle, was employed in this research to detect various objects using the faster R-CNN algorithm as well as YOLOv3. The researchers have presented a comprehensive perspective on the identification of individual face masks and the monitoring of violations of social distancing. The enforcement of face-mask usage and adherence to social distancing guidelines are two crucial measures for mitigating the transmission of COVID-19. The proposed research consists of three main components. Firstly, it involves classifying face masks using a weighted naive Bayes classifier (WNBC) and a deep learning-based, faster region-based convolutional neural network (Faster R-CNN) optimized AGSO to identify appropriate and incorrect face-mask wear in UAV images. Secondly, it focuses on tracking social distancing through object recognition and traceability strategies. This involves real-time recognition of individuals using bounding boxes. During our face mask recognition comparison, the faster R-CNN model outperformed the Single Shot Detector, GravNet, and DGCNN models in terms of speed and accuracy.
The results showed that the face cover with mask achieved a training precision of 98.9%, a recall of 85.6%, and an F1 score of 91%. On the other hand, the Face Without Mask achieved a training precision of 86.9%, a recall of 98.4%, and an F1 score of 92.43%. It is important to highlight that the validation error for the entire collection of 16000 images remains the same or increases, while the training error increases from 12000 to 2000 images. The term for this problem appears to be overfitting. After 12000 training examples, the alignment between the two entities appears to be flawless as they are both following a congruent trajectory. The validation loss initially decreases to a low level as the number of training photos increases, but once more than 14,000 images are used for training, the validation error starts to increase rapidly. Consequently, the effects of social distancing were evaluated on different sets of videos, with crowded areas specifically taken into account to enhance the difficulty of recognition. Moreover, the article presents an expedited R-CNN methodology and demonstrates that object recognition is significantly faster compared to other established, sophisticated techniques. The validation loss, validation accuracy, precision, and recall of the proposed approach have been determined to be optimal.
.
Data availability
The datasets used and/or analysed during the current study available from the corresponding author on reasonable request.
References
Pandey, Jay Kumar, Rituraj Jain, R. Dilip, Makhan Kumbhkar, Sushma Jaiswal, Binay Kumar Pandey, Ankur Gupta, and Digvijay Pandey. "Investigating role of iot in the development of smart application for security enhancement." In IoT Based Smart Applications, pp. 219-243. Cham: Springer International Publishing, 2022.
Rustam, F. et al. COVID-19 Future Forecasting Using Supervised Machine Learning Models. IEEE Access. (2020).
Chamola, V., Hassija, V., Gupta, V. & Guizani, M. A Comprehensive Review of the COVID-19 pandemic and the role of IoT, drones, AI, Blockchain, and 5G in managing its impact. IEEE Access. 8, 90225–90265 (2020).
Gupta, A., Srivastava, A., Anand, R. & Tomažič, T. Business Application Analytics and the internet of things: the connecting link, in: New Age Analytics, Gulshan Shrivastava, Sheng-Lung Peng, Himani Bansal, (eds Sharma, K. & Sharma, M.) 249–273, Apple Academic Press, Canada,. (2020).
Gupta, A., Srivastava, A. & Anand, R. Cost-effective Smart Home automation using internet of things. J. Communication Eng. Syst. 9 (2), 1–6 (2019).
Kumar Pandey, B. et al. Encryption and steganography-based text extraction in IoT using the EWCTS optimizer. Imaging Sci. J. 69 (1–4), 38–56 (2021).
Kirubasri, G. et al. Software-defined networking-based ad hoc networks routing protocols. In Software Defined Networking for Ad Hoc Networks (95–123). Cham: Springer International Publishing. (2022).
Nag, A., Nikhilendra, J. N. & Kalmath, M. IOT Based Door Access Control Using Face Recognition. In 2018 3rd International Conference for Convergence in Technology (I2CT) (pp. 1–3). IEEE. (2018), April.
Babu, S. Z. D. et al. Analysation of Big Data in Smart Healthcare. In Artificial Intelligence on Medical Data: Proceedings of International Symposium, ISCMM 2021 (pp. 243–251). Singapore: Springer Nature Singapore. (2022), July.
Danelljan, M., Hager, G., Shahbaz Khan, F. & Felsberg, M. Learning spatially regularized correlation filters for visual tracking. In Proceedings of the IEEE international conference on computer vision (pp. 4310–4318). (2015).
Ren, S., He, K., Girshick, R. & Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. In Advances in neural information processing systems (pp. 91–99). (2015).
Singh, H. et al. Effective Overview of Different ML Models Used for Prediction of COVID-19 Patients. In Artificial Intelligence on Medical Data: Proceedings of International Symposium, ISCMM 2021 (pp. 185–192). Singapore: Springer Nature Singapore. (2022), July.
Pramanik, S. et al. An overview of IoT privacy and security in smart cities. In AIP Conference Proceedings (Vol. 2495, No. 1). AIP Publishing. (2023), October.
Takahashi, N., Gygli, M., Pfister, B. & Van Gool, L. Deep convolutional neural networks and data augmentation for acoustic event detection. arXiv preprint arXiv:1604.07160. (2016).
Teboulbi, S., Messaoud, S., Hajjaji, M. A. & Mtibaa, A. Real-Time Implementation of AI-Based Face Mask Detection and Social Distancing Measuring System for COVID-19 Prevention. Scientific Programming, 2021. (2021).
Uddin, M. I., Shah, S. A. A. & Al-Khasawneh, M. A. A novel deep convolutional neural network model to monitor people following guidelines to avoid COVID-19. Journal of Sensors, 2020. (2020).
K. Simony an and A. Zisserman. Deep convolutional networks for large-scale image recognition. ArXiv preprint arXiv: 1409.1556, (2014).
Pandey, B. K., Pandey, D., Wairya, S. & Agarwal, G. An advanced morphological component analysis, steganography, and deep learning-based system to transmit secure textual data. Int. J. Distrib. Artif. Intell. (IJDAI). 13 (2), 40–62 (2021).
Pandey, B. K. et al. Effective and secure transmission of health information using advanced morphological component analysis and image hiding. In Artificial Intelligence on Medical Data: Proceedings of International Symposium, ISCMM 2021 (pp. 223–230). Singapore: Springer Nature Singapore. (2022), July.
The lancet digital health, 1(6), e271-e297.
Petrović, N. & Kocić, Đ. IoT-based System for COVID-19 Indoor Safety Monitoring. preprint), IcETRAN, 2020, 1–6. (2020).
QIN, B. & LI, D. Identifying facemask-wearing Condition using image super-resolution with classification network to Prevent COVID-19. Sensors. 20 (18). https://doi.org/10.3390/s20185236 (2020).
Rezaei, M. & Azarmi, M. DeepSOCIAL: Social distancing monitoring and infection risk Assessment in COVID-19 pandemic. Appl. Sci. 10 (21), 7514 (2020).
Yadav, S. Deep learning based Safe Social Distancing and Face Mask Detection in Public Areas for COVID-19 Safety guidelines Adherence. Int. J. Res. Appl. Sci. Eng. Technol. 8 (7), 1368–1375. https://doi.org/10.22214/ijraset.2020.30560 (2020).
Ahmed, I., Ahmad, A. & Jeon, G. An IoT based deep learning framework for early assessment of Covid-19. IEEE Internet Things J. https://doi.org/10.1109/JIOT.2020.3034074 (2020).
Rao, T. V. N., Sai, G. V., Chenna, M. & Reddy, M. S. An application for the detection of COVID-19 face Mask using OpenCV. High. Technol. Lett. 26 (7), 965–975 (2020).
Ndiaye, M. et al. IoT in the wake of COVID-19: a Survey on contributions, challenges and Evolution. IEEE Access. 8, 186821–186839 (2020).
Kumar, S. S. Emerging Technologies and Sensors That Can Be Used During the COVID-19 Pandemic. In 2020 International Conference on UK-China Emerging Technologies (UCET) (pp. 1–4). IEEE. doi: (2020)., August https://doi.org/10.1109/UCET51115.2020.9205424
Sengupta, K. & Srivastava, P. R. HRNET: AI on Edge for mask detection and social distancing. arXiv preprint arXiv:2111.15208. (2021).
Redmon, J., Divvala, S., Girshick, R. & Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 779–788). (2016).
Pandey, D. et al. An approach for object tracking, categorization, and autopilot guidance for passive homing missiles. Aerosp. Syst. 5 (4), 553–566 (2022).
Pandey, B. K. & Pandey, D. Parametric optimization and prediction of enhanced thermoelectric performance in co-doped CaMnO3 using response surface methodology and neural network. J. Mater. Sci.: Mater. Electron. 34 (21), 1589 (2023).
Han, S., Mao, H. & Dally, W. J. Deep compression: Compressing deep neural networks with pruning, trained quantization and huffman coding. arXiv preprint arXiv:1510.00149. (2015).
Han, S., Pool, J., Tran, J. & Dally, W. Learning both weights and connections for efficient neural network. In Advances in neural information processing systems (pp. 1135–1143). (2015).
Simonyan, K. & Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556. (2014).
Henriques, J. F., Caseiro, R., Martins, P. & Batista, J. High-speed tracking with kernelized correlation filters. IEEE Trans. Pattern Anal. Mach. Intell. 37 (3), 583–596 (2014).
Liao, S., Jain, A. K. & Li, S. Z. A fast and accurate unconstrained face detector. IEEE Trans. Pattern Anal. Mach. Intell. 38 (2), 211–223 (2015).
Koo, Y., You, C. & Kim, S. OpenCL-darknet: An OpenCL implementation for object detection. In 2018 IEEE International Conference on Big Data and Smart Computing (BigComp) (pp. 631–634). IEEE. (2018), January.
Kristan, M., Matas, J., Leonardis, A., Felsberg, M., Cehovin, L., Fernandez, G., …Pflugfelder, R. (2015). The visual object tracking vot2015 challenge results. In Proceedings of the IEEE international conference on computer vision workshops (pp. 1–23).
Pandey, B. K., Pandey, D. & Agarwal, A. Encrypted information transmission by enhanced steganography and image transformation. Int. J. Distrib. Artif. Intell. (IJDAI). 14 (1), 1–14 (2022).
Pandey, D., Pandey, B. K. & Wairya, S. Hybrid deep neural network with adaptive galactic swarm optimization for text extraction from scene images. Soft. Comput. 25, 1563–1580 (2021).
Xu, X. & Kakadiaris, I. A. Joint head pose estimation and face alignment framework using global and local CNN features. In 2017 12th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2017) (pp. 642–649). IEEE. (2017), May.
Deepa, R., Anand, R., Pandey, D., Pandey, B. K. & Karki, B. Comprehensive performance analysis of classifiers in diagnosis of epilepsy. Mathematical Problems in Engineering, 2022. (2022).
J. Redmon, S. Divvala, R. Girshick and A. Farhadi, "You Only Look Once: Unified, Real-Time Object Detection," 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 2016, pp. 779-788, https://doi.org/10.1109/CVPR.2016.91.
Bessant, Y.R.A., Jency, J.G., Sagayam, K.M. et al. Improved parallel matrix multiplication using Strassen and Urdhvatiryagbhyam method. CCF Trans. HPC 5, 102–115 (2023). https://doi.org/10.1007/s42514-023-00149-9.
Iandola, F. N. et al. SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and < 0.5 MB model size. arXiv preprint arXiv:1602.07360. (2016).
Boopathi, S., Pandey, B. K. & Pandey, D. Advances in Artificial Intelligence for Image Processing: techniques, applications, and optimization. In Handbook of Research on Thrust Technologies’ Effect on Image Processing (73–95). IGI Global. (2023).
He, K., Zhang, X., Ren, S. & Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 770–778). (2016).
Pandey, B. K., Pandey, D., Wariya, S. & Agarwal, G. A deep neural network-based approach for extracting textual images from deteriorate images. EAI Endorsed Trans. Industrial Networks Intell. Syst. 8 (28), e3–e3 (2021).
Nam, H. & Han, B. Learning multi-domain convolutional neural networks for visual tracking. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 4293–4302). (2016).
Ranjan, R., Patel, V. M. & Chellappa, R. Hyperface: A deep multi-task learning framework for face detection, landmark localization, pose estimation, and gender recognition. IEEE Trans. Pattern Anal. Mach. Intell. 41 (1), 121–135 (2017).
Ryu, S. K., Kim, T. & Kim, Y. R. Image-based pothole detection system for ITS service and road management system. Mathematical Problems in Engineering, 2015. (2015).
Qasim, S. R., Kieseler, J., Iiyama, Y. & Pierini, M. Learning representations of irregular particle-detector geometry with distance-weighted graph networks. Eur. Phys. J. C. 79 (7), 1–11 (2019).
Anand, R. et al. Hybrid convolutional neural network (CNN) for Kennedy Space Center hyperspectral image. Aerosp. Syst. 6 (1), 71–78 (2023).
Pandey, D. & Pandey, B. K. An efficient deep neural network with adaptive galactic swarm optimization for complex image text extraction. In Process Mining Techniques for Pattern Recognition (121–137). CRC. (2022).
https://github.com/chandrikadeb7/Face-Mask-Detection/tree/master/dataset
Acknowledgements
The authors would like to express gratitude to Department of Statistics, College of Natural and Computational Science, Mizan-Tepi University, Tepi, Ethiopia.
Funding
The author(s) received no financial support for the research, authorship, and/or publication of this article.
Author information
Authors and Affiliations
Contributions
Binay Kumar Pandey and Digvijay Pandey perform experiment, wrote manuscript. Mesfin Esayas Lelisho draw figure, wrote manuscript. All authors approve the final manuscript.
Corresponding authors
Ethics declarations
Conflict of interest
The authors declare that they have ‘no known conflict of interests or personal relationships’ that could have appeared to influence the work reported in this paper.
Ethics approval
Not Applicable (as the results of studies do not involve any human or animal).
Consent to participate
Not Applicable (as the results of studies do not involve any human or animal).
Consent for publication
Not Applicable (as the results of studies does not involve any human or animal).
Additional information
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Pandey, B.K., Pandey, D. & Lelisho, M.E. Face mask identification with enhanced cuckoo optimization and deep learning-based faster regional neural network. Sci Rep 14, 29719 (2024). https://doi.org/10.1038/s41598-024-78746-z
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41598-024-78746-z
Keywords
This article is cited by
-
Blockchain based electronic educational document management with role-based access control using machine learning model
Scientific Reports (2025)
-
The impact of urbanization on land use land cover change using geographic information system and remote sensing: a case of Mizan Aman City Southwest Ethiopia
Scientific Reports (2025)
-
Innovative deep learning classifiers for breast cancer detection through hybrid feature extraction techniques
Scientific Reports (2025)
-
Advanced AI-driven emergency response systems for enhanced vehicle and human safety
Iran Journal of Computer Science (2025)
-
Innovative Quantum PlasmoVision-Based Imaging for Real-Time Deepfake Detection
Plasmonics (2025)