Trusted Publishing Platform
Serving Researchers Since 2012

A Survey on Components of an End-to-End Face Detection System: Algorithms, Limitations and Intelligent Computing

DOI : https://doi.org/10.5281/zenodo.18068732
Download Full-Text PDF Cite this Publication

Text Only Version

 

A Survey on Components of an End-to-End Face Detection System: Algorithms, Limitations and Intelligent Computing

Snehal D. Patil

Electronics and Telecommunication COEP Tech. University, Pune, India

Dr. Prashant P. Bartakke

Electronics and Telecommunication COEP Tech. University, Pune, India

Mr. Rajesh Chavan

Research and Development Fourfront Pvt. Ltd, Pune, India

Dr. Mukul S. Sutaone

Electronics and Telecommunication IIIT Allahabad, Allahabad, India

Abstract – Face detection is the very first step for an efficient face recognition system. The success of an application for face recognition is hinged rigidly with the implementation of a coher- ent face detection system. Before designing an end-to-end system, foremost attention must be provided in choosing a desired face detection technique based on system architecture, complexity, inference time and limitations it offers. Most of the real-world applications like home automation, attendance, tourism, banking, security, automobile, immigration, retail, healthcare expects a prompt response, making it essential that inference time from each module of an application should be optimized. Face detec- tion is constrained by several challenges that consists of occlusion, illumination, fake face, scale, pose, low resolution, make-up, extreme expressions, reflection, complexion etc. The emergence of cloud and edge computing platforms enabled to meet the huge deficit among training data and computation resources. The deployment of the model on embedded platform is often a concern due to enormous model size. In this survey exhaustive analysis of various face detection techniques is put through regarding network architecture, limitations, python packages, performance metrics and advantages. Also, the components of an end-to- end application development imbibing face detection is accom- plished. Custom face detection models hunt effectively for the existing constraints and uncontrolled conditions. Integration of face detection systems and Advanced Driver Assistance Systems (ADAS) platforms on cloud instances is consummated to cater for emerging need of security in automotive sector. Towards the end, a brief overview of the challenges in this field and dimensions for future research are contemplated.

face detection, driver monitoring system, intelligent com- puting, cloud instances, PYPI packages

  1. INTRODUCTION

    Face Localization is the foundation for computer vision tasks. Intelligent computing is the buzz word in advanced driver assistance systems (ADAS), surveillance and driver monitoring systems (DMS). All these systems comprise of a video capturing camera and an analysis system to initiate a proactive action for prevention of hazard. Deep Learning is a

    boon sustaining scalability of the data collection devices and enhancing the computing abilities to take a decision accurately. Face Detection is affected due to different factors such as pro- file view, styling accessories, fake images, scale, and cluttered background. Face Localization is the specific case of object detection task. The entire journey towards making life simpler started with the emergence of technologies imitating human thinking. The advanced computing techniques are modifying the dimensions in which data can be interpreted and analyzed. The inter relation among the above techniques and the way these techniques handle data is depicted (refer Figure 1). Artificial intelligence is the capability of computers to acquire information from the input data. It solves problems effectively by devising optimal and adaptive techniques without human intervention. The origin of Intelligent systems can be traced back with a novel question Can machines think? [1]. If the evaluator is unable to distinguish the replies between the person and a general-purpose computer in an imitation game, the processor is declared to be the winner. The goal is straight forward to persuade the assessor about their interaction with a human being rather than an intelligent device. Machine Learning has been introduced to make our life easier. Early intelligent systems used programmed conditional statements to analyze the user information. But with Machine Learning, data is made available to learn discriminative features and understand patterns from it. Machine Learning detour the need to undergo repetitive coding for every new query encountered unlike accustomed issues, the similar method needs to be used with different dataset. Deep Learning utilizes computational techniques and experiential data for training purposes inspired by our brains own network of neurons. It interprets the information that acts as an input and process the information from cumulative experiences. The base line for Deep Learning algorithms derives from experience gathering and active learn-

    ing. Computers learn through a network of neurons and boost their properties without specific programming or mathematical modeling.

    Deep Learning is gaining significant attention in various application domains. The detailed study of face detections commences with the conventional methods, moving towards machine learning approaches and settles down on deep learn- ing techniques. Processing enormous amounts of data imposes restrictions on computing resources. To surpass this challenge, cloud and edge computing proves effective. Comprehensive discussion of different cloud instances and machine learning ADAS platforms is performed. Cloud instances such as AWS Sagemaker, Azure machine learning, IBM SoftLayer, and Google Cloud AutoML empower model development and de- ployment allowing cross framework interoperability and GUI development tools. NVDIA, NXP, and AWS are promoting the design of driver monitoring systems featuring multi framework and inference engine support. System architecture, parallel computing, hardware, and software acceleration are the key aspects affecting the performance of programming devices. The first step towards an application development is selec- tion of the system model with consideration of model size, computing resources, training, and inference time. Graphic processing unit (GPU) is an alternative solution to quench the limited processing abilities of a general-purpose computer during training. Compute Unified Device Architecture from NVIDIA is a parallel processing GPU platform enabling computation intensive tasks of an application to be dispatched concurrently. Field programmable gate array (FPGA) has an added advantage of hardware reconfiguration and model optimization. To enable real time inferencing, deployment of trained model on embedded platform is significant. The model configuration is gigantic making direct flashing on end device less feasible. The amalgamation of various components lays down firm pillars for a competitive application. This survey provides literature in emerging intelligent computing facilities, python packages, network architecture, advantages and flaws, factors for inaccurate detection, loss functions, performance metrics, dataset for face detection application, cloud computing platforms, machine learning based advanced driver assistance systems. For the organization of the paper and subsections corresponding to it (refer Figure 2)

  2. EVOLUTION OF FACE DETECTION METHODS
    1. Image Processing face detection methods

      Face processing is gaining attention in the arena of computer vision domain owing to uniqueness of an individual face. Hu- man face detection is a significant research area incorporating diverse applications like video surveillance, access contro, face verification, face expression, advanced computer, and human interaction. Early face localization can be roughly cate- gorized into local facial features detection, template matching and image invariants. Conventional face detection systems originally were based on facial feature excerption utilizing low level computer vision methods and classification based

      on statistical models [2] [3] [4] [5]. Template matching com- prising several correlation templates are used to detect local sub features [6]. Wavelet packet decomposition encompassing skin colour sieving and likelihood categorization of facial textures was widely used for rapid face detection. In case of image invariants, there is an assumption of spatial image relationships unique to all image patterns and under various imaging conditions [7]. The early methods have drawback of limited global constraints applied on face templates and the features extracted are greatly affected by noise or expression or viewpoint. Correlation based template matching methods are computationally expensive and require a large amount of storage. Alternative approaches for human face detection instead of handcrafted features are based on neural networks wherein [8] [9] ample information is provided to learn and identify patterns in data.

    2. Machine Learning face detection methods

      In these methods handcrafted features are extracted. Al- though the methods are popular and widely used, it has limitations in terms of computation time and detection of false positives.

      1. Haar Cascade or Viola Jones: Several techniques are proposed and widely used for Face Detection, but the foremost and successful algorithm widely used was in 2001 by Viola and Jones. They proposed a framework for object detection in real time from video footage. This algorithm was proposed long back when Deep Learning had not been even given thought off. There are two approaches for seeking information from an image feature based and pixel based. The advantage of using feature-based systems is that theyre faster as compared to pixel-based systems. In the above figure, the darker areas represent pixel value 1 and lighter areas represent pixel value

        0. Haar features are used for finding features in an image. Feature extraction is dominated by sharp intensity variations across a line or any other arrangement. In order to extract vertical features with low intensity pixels on the right side and high intensity ones on the left side (refer Figure 3A) Also, to procure horizontal features with low intensity pixels above the edge and low intensity pixels below it (refer Figure 3B). The purpose is to calculate the sum of all pixel intensities in the darker area and the sum of all pixel intensities in the lighter area of the haar feature. Now if there is an edge appearing between darker and lighter pixels, then feature value is evaluated approximately near to 1. This mathematical process is carried out to synthesize different features from image or video under consideration.

      2. Histogram of Oriented Gradients (HoG): It is robust as compared to haar cascades classifier and works effectively in different illumination conditions. For workflow of HoG (refer Figure 4). This technique is based on counting the gradients in a localized portion of an image. Feature detection principle works on transforming from image space to parameter space. Image space consists of Cartesian coordinates as parameter while parameter space consists of slope and y intercept as parameter. So, point in image space maps to line in parameter

        Fig. 1. Inter Relation among intelligent computing systems- artificial intelligence, machine learning, and deep learning

        Fig. 2. Structure of the Survey

        for this is to reconfigure the parameter space in terms of distance from origin and angle. As both distance and angle are finite, the size of accumulator reduces drastically. The calculated gradients are further processed to evaluate the his- tograms that are further converted to HoG description vectors and classified with the aid of Support vector machine (SVM) classifier. Viola Jones is a preliminary machine learning ap-

        Fig. 3. Haar Feature Extractor

        space. As the value of slope ranges from minus infinity to plus infinity, there is limitation that parameter space requires large size of accumulator and memory computations. The solution

        Fig. 4. Architecture of Histogram of Oriented Gradients

        proach for rapid processing of images with high detection

        accuracy indexing huge amount of video and image data. The three major contributions that distinguish this methodology from others are Integral Image, small critical visual features with Adaboost learning algorithm, combination of complex classifiers in cascade focusing on promising object regions and eliminating background regions [10]. Deciphering the ingredients of first ever real time face detection system was illustrated in [11]. Multiple redundant detections from Viola- Jones algorithm is conquered with a robustness argument in the past processing step. Speed up Robust Features (SURF) [12] is an extension work of approach yielding much faster training convergence with AUC as the single convergence criterion as against Viola Jones framework which utilizes two inconsistent metrics (sensitivity and discerning rate). Human facial factors such as eyes, nose, mouth, and face are detected with haar cascade object detector [13]. Li et al. [14] proposed a system of haar cascades with three additional classifiers to get rid of non-human faces. An improved Haar cascades algorithm for face localization with Microsoft HoloLens achieved 12% on average higher detection efficiency and four-fold detection speed than that of existing approach [15]. HOG with SVM utilizes a classifier that constructs high dimensional concate- nated HOG features. The high dimensional representation not only results in time consuming training procedures but also leads to slow detection speed. Human detection system based on HOG is implemented in [16] [17] [18]. HOG can perform better when learned via MMOD (max margin object detection) [19]. Partially overlapping windows with objects are difficult to imbibe into training set as these are neither a false alarm nor true detection. MMOD leads to convex optimization by working on missed detections and false alarms. The next logical step after an efficient face detection model is human detection. Local Binary Pattern LBP [20] has shown effec- tiveness in diverse tasks of face recognition, facial detection, facial expression analysis and demographic classification. LBP is a non-parametric descriptor summarizing the local structures in an image by comparing each pixel with its neighboring one. Conventional methods accomplished significant detection performance on existing approaches prevelant at that time

        [9] [8]. Face localization encounters two major challenges: a.

        Enormous visual changes in bestrewed backgrounds b. huge search area for diverse face sizes. These challenges impose a time efficiency requirement and effective binary classifier. Deep Learning significantly improves training time, resources utilization and accuracy.

    3. Deep Learning Based Methods

    Deep Learning approach handles enormous amount of data for taking appropriate decision. The tradeoff between accuracy and computation needs to be managed with utmost care. The architecture of SSD is as shown in (refer Figure 5). Table I shows comparison of different SSD techniques. Architecture of CNN is shown (refer Figure 6). Comparison of Deep learning- based CNN models are shown in Table II. The representation of YOLOV3 architecture is shown in (refer Figure 7). Table III shows a comparison of different YOLO techniques. Table

    IV represents OPENCV frameworks for face detection. Figure 8 demonstrates MTCNN architecture.

    1. Single Shot Detector (SSD): SSD handles detections with two uilding blocks, a backbone, and head. Backbone extracts features from pre-trained ResNet model excluding fully convolutional layers and trained on benching dataset such as ImageNet. The SSD head comprises of convolutional layers following the backbone and the outputs consists of confinement boxes and objectness score in the spatial domain. To detect objects, the entire image is divided into grids. Anchor boxes enable detecting an object of a specific size and shape within the grid cell. Cricket ground maps to the vertically narrow anchor box while sky scrapper relates to horizontally thin one. The pre-defined cells are compared with ground truth for evaluation of intersection over union for estimating bounding box coordinates and object class. Zoom parameters determine the level to which the cell must be scaled to fit faces of diverse sizes for detection. A 4*4 grid, 2*2 grid and 1*1 grid detect smaller objects, mid-sized objects and the objects that cover the entire image respectively. Mediapipe Face detection supports multiple face detection along with key point detection, iris detection, face mesh detection and hair segmentation. It also supports pose estimation, object detec- tion, and object tracking. Blaze face is a lightweight algorithm working effectively on mobile GPU enabled anchors.

      SSD discretizes the image space into a set of default boxes with aspect ratios and scales [21]. At inference time, the system produces an objectness score for the existence of object in each default box and performs adjustments to better fit the object shape. Also, the system integrates estimations from multiple feature maps with varying resolutions to deal with objects of different sizes.

      Feature Agglomeration Networks [22] are motivated by feature pyramid networks (FPN). The prime objective is to uti- lize multiscale features for aggregating higher level semantic feature maps to boost lower-level feature maps via hierarchical agglomeration.

      To deal with multiscale faces, one way is to train multishot single scale detectors by using image pyramid to train multiple separate single scale detectors each for one specific scale. However, this is computationally expensive since it must pass through the network multiple times during testing. Another approach is to train a single shot multiscale detector by exploiting multiscale feature representations requiring only a single pass through the network while testing. Single shot scale invariant detector (S3FD) [23] contributes to face localization in the following three aspects 1. Scale equitable detection structure for multiscale face detection 2. Scale compensation anchor matching strategy for tiny face detection 3. Reducing false alarms with max out background label. It is common observation that anchor based detectors fail to detect tiny faces. Max out operation injects local optimal solution to deal with unbalanced binary classification problem of negative anchors (i.e. background) and only few positive anchors (i.e. face). This happens due to dense tile of small anchors, contributing to most of false positive faces.

      TABLE I

      Comparison of SSD based face detection methods. Algorithms are evaluated in terms of feature extractor, dataset, performance metrics, loss function, training device, and frames per second(FPS).

      Sr No Algorithm Name Feature Dataset Performance Loss

      Training Device FPS

      Extractor metric function

      1 REF[21] SSD Convolution

      filter

      2 REF[22] FANet VGG-16
      3 REF[23] S3FD VGG-16
      4 REF[24] SSH VGG-16
      5 REF[25] DSFD VGG-16
      6 REF[26] Pelee Peleenet
      7 REF[27] Pyramid Box VGG-16

       

      8 REF[28] RefineDet VGG-16

      Resnet-101

      9 REF[29] Fcos Resnet-101 Resnet-50
      10 REF[30] LFFD
      11 REF[31] SANet Resnet-50
      12 REF[32] HAMBox Resnet-50

       

       

      Fig. 5. The Architecture of SSD model

      Single stage headless (SSH) [24] performs far better than state of art methods by removing fully connected layers (head) of VGG-16 backbone. SSH can detect faces at various scales without generating an image pyramid. SSH uses simple convolution to achieve larger window effect. During training phase anchors are assigned to three modules M1, M2 and M3 depending on face size. One unique aspect of this framework is that an anchor is assigned to ground truth face if and only if it has higher IoU than 0.5.

      Dual shot face detector (DSFD) [25] is a variant of single stage detector comprising of two stream design networks. It is an efficient face detection framework consisting of complimentary modules such as feature enhancement module (FEM), progressive anchor loss (PAL) and improved anchor matching (IAM). FEM ensures discriminability and robustness of features, PAL combines hierarchical loss and pyramid anchor that assigns smaller anchor sizes in first slot and larger sizes in second shot, IAM uses anchor partition strategy and anchor-based data augmentation to better match anchors and ground truth faces for effective regressor initialization.

      Pyramid Box [27] address the challenge of unconstrained face detection. To deal with hard face localization contex- tual data is exploited with context anchor, low level feature pyramid network and context sensitive structure. Multiscale training samples are generated from data anchor sampling to ensure diversity of data for tiny faces. Devised Receptive Field Block net (RFB-net) [33] motivated by the receptive fields (RF) in human beings to extract discriminable and

      robust features. Pelee [26] is another efficient single stage detector that handles tradeoff between computer power mem- ory resources [28] inherits one stage approach consisting of anchors and object detection network to adjust size of anchors. Features from anchor refinement module are forwarded to object detection network via a transfer connection block. Blaze face [29] is a novel light weight face detector featuring super embedded devices of 200-1000+ fps on flagship embedded devices based on mobile level v1/v2. Fully convolutional one stage object detection (Fcos) [30] performs pixel object detection analogous to sematic segmentation. Fcos is anchor box free as well as propsal free. Fcos meticulously avoids hyper parameters and complex computation concerned with anchor boxes. Anchor based detection is sensitive to scale, aspect ratio and location of anchor boxes. Low quality anchor boxes are discarded by evaluation of centerness of a location. At inference time, the classification scores get down-valued when multiplied by center-ness to reject low quality predicted bounding boxes. Light and fast face detector (LFFD) [31] enables deployment on edge devices utilizing receptive field and extended receptive field. Anchor based unable to cover all face scale, threshold for IOU is empirically set, sample imbalance and redundant computation.

      For large faces (RF), for medium, faces (ERF with little context) for small faces (ERF with relevant context). Pyramid Box ++ [32] is extension work based on pyramid box boost- ing face detection with balanced data anchor sampling, dual pyramid anchors and dense context module.

      Fig. 6. Architecture of Convolutional Neural Network (CNN)

      Integration of multiscale features can contribute significant noise. In smooth attention network (SANet) [34] attention guided feature fusion module [AFFM] and smoothed context enhancement module (SCEM) manages the griding artifacts. AFFM performs attention wise feature fusion artifacts. AFFM performs attention wise feature fusion of high and low level to reduce noise. SCEM manages the griding artifacts contributed by deconvolutional layer to preserve local spatial information.

    2. Convolutional Neural Networks (CNN): Convolutional neural networks are the most successful and commonly used algorithms in computer vision tasks. CNN basically consists of three parts convolution layers: it consists of kernels of varying sizes and are used for feature extraction, non-linear layers consist of an activation function that deals with the nonlinear functions, pooling layer to get statistical information about the neighboring pixels. The convolution layer typically consists of kernels (convolutional filters) sliding across the dimensions of an image. Feature extraction is a mathematical operation performing dot product between corresponding values of the image and the convolutional filter. The main advantage in using CNN is that the receptive fields share the same kernels, minimizing the memory constraints as compared to deep neural networks. The main task of the pooling layer is sub- sampling and gathering local statistical information from the feature maps. Equivalent representation, sparse representation and parameter sharing are three key benefits of CNN. [35] is the first to show that CNN can yield better perfor- mance on Pascal VOC as compared to the system based on

      HOG. The proposed system has three subsections comprising of category independent region proposals utilizing selective search, large CNN for fixed length feature extraction and set of SVMs. During inferencing 2000 category independent region proposals are generated from input image. [36] pointed out the trade-off between strong discriminative features and efficient computation. To promptly reject background regions and improve localization capabilities, a calibration stage is introduced after detection stage. [37] [38] showcased that regional proposals are computationally intensive.

      RPN (Region proposal network) is a fully convolution network estimating object boundary and objectness confidence simultaneously thereby enabling faster inferencing. [39] [40] introduced attribute aware face detection, Faceness-Net. Part- ness map relates the presence of specific facial component in the image. [41] developed contextual multiscale R-CNN (CMSRCNN) by additionally providing context information to identify difference between real faces with bodies and fake face without bodies. Here multiscale is rooted both in region proposal and ROI layer to deal with tiny faces. Also, contextual reasoning is added for challenging face detection.

      Lower layer features correspond to edges and corners per- taining to localization information. Deeper layer features are class dependent contributing to high end tasks of face detec- tion. [42] demonstrated the role intermediate layers features called hyper features for training of different tasks under consideration. Features common to tasks can be combined with feature fusion technique translating the features to a common

      subspace by linear or non- linear combination.

      A Supervised Transformer Network [43] employing a cas- caded CNN for RPN to predict face regions and correspond- ing landmarks. Facial landmarks are warped with candidate regions to obtain canonical positions of face. Finally, RCNN verifies the candidate regions for valid or invalid face.

      Detection accuracy is improved with the development of several techniques including position sensitive average pool- ing, Multiscale training and testing and on-line hard example mining strategy in R-FCN [44]. R-FCN improved detection by introducing additional small anchors and modified the position sensitive ROI pooling to a smaller size for tiny face detection. Position sensitive average pooling was used instead of normal average pooling for last layer for enhance embedding. In R-FCN the feature maps are more expressive as unnaturally injecting fully connected layers into Resnet is avoided and easier learning of class score and bounding box is accomplished.

      Conventional CNN simply consists of stack of filter lay- ers where input passes through all of them before reaching classifier. It is the well-known fact that deeper layer possess discriminative capabilities and lower layer rejects non-face samples.

      [45] proposes Inside cascaded structure that introduces face/ non-face classifiers at different layers within same CNN. To determine which samples should be passed to the data routing layer. Early rejection classifier (ERC) predicts face probability to determine which samples should be passed to data routing layer. Data Routing is a mechanism where different layers are trained via different samples, deeper layers dealing with more difficult Samples. Along with this, contextual CNN boosts the detection accuracy by usage of body part information.

      In [46] presents Scaleface network that does not require image pyramid (IP) having moderate complexity. The design of appropriate receptive fields is essential for multiscale face detection. Recent methods can be categorized into two classes: Scale Variant based method and Scale Invariant based meth- ods. The [47] proposes an improved faster RCNN framework by combining various techniques including features concate- nation, hard negative mining, Multiscale training, model pre- training and proper calibration of key parameters. A novel ap- proach in [48] derives robust face detection methodology face RCNN based on faster RCNN. One key technique considered here is bootstrapping with online hard example mining OHEM. The key idea is to collect hard samples and feed them again to the network to strengthen discriminative power. The loss function represents how effectively the network performs, the generated proposals are sorted by their losses and top N worst performing examples are taken as hard examples.

      Face Boxes is a powerful lightweight network structure comprising of Rapidly Digested convolutional layers (RDCL) and multiscale convolutional layers (MSCL)[49]. RDCL en- ables real time face detection and MSCL handles faces over various scales. RDCL is designed to quickly reduce input spa- tial size by suitable kernel. The anchors of RPN are associated with last convolutional layers whose features and resolution

      are too weak for handling faces of diverse sizes. Also, anchor associated layer detects face within corresponding range of scales and has single receptive field that cannot cope up with different scales. MSCL takes care of problems related to RPN by discretizing anchors over multiple layers with different resolution to handle faces of various sizes.

      DSFD (Different Scale Face Detector) [50] technique han- dles small,and scaled faces. Feature maps for small faces shrink gradually over the depth of convolutional neural net- work and can hardly detect faces less than 15×15 pixel.

    3. You Only Look Once (YOLO): Conventional classifiers are modified and remodeled to perform localization. The algo- rithm applies the model at multiple locations and dimensions. Candidates with high scores are treated as detections. Yolo Face is variant derived from Yolov3.Yolo performs single stage detections by dividing the image into grids and estimating the bounding boxes with predicted probabilities. The model glances at the entire image during test time, so predictions are guided by global context. Yolo divides the image into grids with centerpoints. The number of grid sizes range from 3*3, 4*4 and 16*16, there is no fixed rule for number of grids. Each grid is looked for face detection only at one time in a forward pass. Therefore, it is called You Look Only Once. For the summary of YOLO development stages (refer Figure 9).

      Non maximum suppression is entrenched on evaluating the maximum overlap ensuing a unique bounding box. Yolo utilizes CNN layers with stride two as against pooling layers capturing low level features to detect small objects. The earlier approach for object detection is based on classifiers and objectness score. In YOLO a single stage framework estimates the bounding boxes and class probabilities for an object under consideration. Here, object detection is formulated as a regression problem unlike classification approach consid- ered earlier. The development of YOLO series is analyzed across depth and breadth in [51] [52]. The YOLO versions for face detection are designed in [53] [54] [55]. YOLO can be efficiently implemented on FPGA benefiting from its computation capabilities [56] [57]. YOLO-LITE [58] enables the deployment of YOLO algorithm on portable devices.

    4. Multitask Cascaded Convolutional Neural Network (MTCNN): MTCNN consists of three CNN assigned a distinct role and successor refines the detections. Stage-I: P-net scales the picture to enable detections for diverse face range. Con- volution operation with 12*12 kernel progressively traverses through all scaled pictures detecting for faces and respective locations. Multiple redundant detections can be mitigated with the application of non-maximum suppression. NMS criteria may be a large bounding box or large confidence. Stage-II: R- net If the bounding box is out of bounds, the portion of image inside bounding box is copied to new array and remaining everything is filled with a zero. R-net handles bounding box extending beyond image boundaries by creating a new array and setting out of bound values to zero. R-net confines the exact face detection bounding boxes eliminating the redundant ones and returning appropriate square shape boxes. R-net output is like P-net, but it includes a new, more accurate one.

      TABLE II

      Comparison of CNN based face detection methods. Algorithms are evaluated in terms of feature extractor, dataset, performance metrics, loss function, and training device.

      Sr No Algorithm Name Feature Extractor Dataset Performance metric Loss function Training Device
      Sr No Algorithm Name Feature Extractor Dataset Performance metric Loss function Training Device
      1

      2

      REF[59]

      REF[60]

      TinaFace

      RetinaFace

      Resnet-50

      Resnet-50

      WiderFace

      WiderFace

      AP-92.4 Focal loss, DIoU crossentropy lo

      AP-91.7 Smooth L1los

      FDDB TPR-0.880
      3 REF[61] Mask RCNN Resnet-101 AFW

      WiderFace

      AP-95.97 Multitask los AP-0.662
      WiderFace Accuracy-91.1
      AFW AP-99.90
      4 REF[62] RefineFace Resnet-50 Pascal Face

      FDDB

      AP-95.45 Scale aware margi TPR-0.9911
      MAFA Accuracy-95.7
      5 REF[50] DSFD VGG-16 FDDB Recall rate-99.22 Least square

       

       

      Fig. 7. Architecture of You Only Look Once (YOLO)

    5. OpenCV Face detection methods: In [69] detections are performed merging OpenCV library with CAFFE and TensorFlow framework. OpenCV performs the face detections with two options.

    Before deployment of an application on embedded platform, the critical factors affecting the performance of the system must be considered. We have presented exhaustive discussion on several deep learning approaches for face detection. The success of any application depends largely on utilizing the advantages and eliminating the disadvantages with software optimization. Table V shows the benefits and drawbacks of various deep learning methods. Also, there exists methods that are not a part of the categories considered above or typical methods for handling constraints related to face detection. Multitask face detection [70] [71] [72], Tiny face detection

    [73] [74], face detection analysis toolkit [75] [76], Occlusion efficient face detection [48] [77] and rotation invariant face detection [78] are popular methods for role specific face localization.
  3. Overview Of Different PYPI Software Package

    Face Detection implementation is reaching new heights with the usage of different python packages. The packages can be installed on computing devices and algorithms can be executed with a simple command line interface.Different software packages for face detection are analyzed in Table

    VI. Figure 10 depicts different face detection software im-

    plementation available. Discussion of detailed architecture of different models and their variants gave a deep insight for underlying principle behind the detection mechanism. Before selecting a particular model for an application, one must be aware of the limitations and benefits each one has, as provided in Table VII. Inaccurate face detection affects the performance of an application.Figure 11 contributes for factors resulting in inefficient face detection based on underlying architecture and performance metrics. The success of the practical computer vision applications is dominated by haar cascade classifiers due to their fast execution time. Several algorithms based on hand crafted features utilizing machine learning ideology emerged for effective face detection. In 2012, a deep learning era emerged contributing to the most accurate and quick face localization techniques. It is crucial to interpret detailed architecture of different models and their variants to get a deep insight into the underlying principle. Custom face detection is essential to overcome the specific challenges related to tiny faces, blurry faces, modality, illuminated faces, rotated faces, extreme expression faces. Custom Face Detection can be implemented with Yolov5, Yolov7, TensorFlow and Detectron models, to surpass the limitations mentioned beforehand. Face Detection can be implemented with either extraction of hand- crafted features or utilizing some deep learning based pre- trained models. Custom face detection models can withstand diverse real time conditions encountered in day-to-day life. To start with, one can go for pre-trained models. If the desired accuracy is not achieved then one can test for custom models.

    TABLE III

    Comparison of YOLO based face detection methods. Algorithms are evaluated in terms of feature extractor, dataset, performance metrics, loss function, and training device.

    Sr No Reference Variant Backbone Neck Head Loss Improvements
    1 REF[63] YOLOv1 GoogleNet FC>7x7x(5+5+50) MSE Directly fit the location of the bounding box
    Batch Normalization,
    High Resolution classifier, Convolutional with anchors,

    2 REF[64]

    YOLOv2/ YOLO9000

    Darknet-19 Conv>13x13x5x(5+20) MSE

    Conv>13x13x5x(5+80)

    Dimension clusters, Direct location prediction, Fine-Grained features, Multiscale training, Hierarchical classification Multiscale prediction,

    3 REF[65] YOLOv3
    4 REF[66] YOLOv4
    5 REF[63] YOLOv5
    6 REF[67] YOLOv6
    7 REF[68] YOLOv7

     

    SPP+PAN Conv>13x13x5x(5+80)

    >13x13x5x(5+80)

    >13x13x5x(5+80)

    Conv>13x13x5x(5+80)

    CIoU Mosiac for data enhancement, Using multi anchors for single groundtruth,

    Eliminate grid sensitivity (sigmoid),

    MSE loss>GIoU loss>CIoU loss Adaptive anchor strategy,

    SPP+PAN >13x13x5x(5+80) GIoU Adopt Focus Structure,
    >13x13x5x(5+80) GIoU CSP Structure More training epochs,
    PAN Decoupled head SIoU

    CIoU

    Self Distillation,

    Gray border of images

     

    Framework Memory Function

    centralized cloud [79]. Cloud computing is a phenomenon of utilizing remotely located servers on the internet to handle end user data. Cloud computing demands low latency, scalability, and privacy [80]. Table VIII presents different cloud instances

    Fig. 9. Evolution of YOLO variants

  4. Overview of Intelligent Computing Techniques

    The information generated from end terminals such as cameras needs to be processed for analysis or forwarded to train a model. Expeditious training and inferencing demand significant computation resources. Training a deep learning model is a complex process involving millions of parameters and high dimension data. Cloud computing leverages the critical computational resource constraint by relocating data to

    instantly to initiate a payment process. Handling this applica-

    tion through cloud will incur overhead in terms of processing and forwarding delay. Addition of a greater number of users can lead to bottleneck for bandwidth sensitive applications. Sending personal data to the cloud raises safety concerns and can violate privacy aspects.

    Edge computing overcame the above challenges by im- plementing the computing nodes close to the source. Edge computing facilitates processing and management of client data at periphery of network in the proximity of the originating source. Edge computing demands a high-end computer node, heterogenous communication and privacy. Google released TensorFlow and TensorFlow Lite framework for implemen- tation on edge devices. Facebook developed Caffe and Caffe2 framework for deployment of deep learning models on Rasp- berry Pi and mobile devices. PyTorch is another framework by Facebook to promote production of research prototypes. NVIDIA GPU powered by CUDA and cuDNN promotes par- allel computing ability on self-developed embedded platforms such as NVIDIA Jetson Nano, NVIDIA Xavier, and NVIDIA Orin. To integrate the model on embedded devices, mod- els with reduced parameters can boost memory and latency

    TABLE V

    Summary of benefits and drawbacks of Deep Learning based Face Detection methods. Benefits enables the end user to choose an appropriate algorithm depending on the resources available. Drawback puts check on the computing devices processing capability

    Reference Method Benefit Drawback

    Single stage computation eliminating proposal generation and feature resampling

    1. REF[21] SSD
    2. REF[38] Faster RCNN
    3. REF[22] FPN
    4. REF[41] CMS-RCNN
    5. REF[24] SSH
    6. REF[30] FCOS
    7. REF[35] R-CNN

      stages.

      Faster than YOLO.

      Faster predictions at different scales using multiple layers achieving high accuracy even at low resolution input.

      Accurate and Robust due to explicit region proposal and pooling.

      Merges low resolution semantically

      strong feature with high resolution semantically weak feature

      Better feature extraction with larger window by duplication of head.

      Boosts detection accuracy by using contextual body part information.

      Simple convolution to achieve larger window effect.

      The anchor is assigned a ground truth face if and only if it has IoU higher than 0.5.

      Face detection at various scales without generating image pyramid.

      Anchor box free and proposal free.

      Avoids some hyperparamters and complex computation concerned with anchor boxes. Enhanced face localization due to region proposals generated by Selective search

      Estimates object boundary and objectness

      Multiscale feature maps alone are used for prediction and thus high resolution semantically weak feature map may fail to perform accurately.

      Increased computation time as proposals is generated in the first stage and classified in second stage.

      The fastest detector operates at only 10s of fps. ROI pool layer builds features from last layer and unable to detect small faces.

      Ignores context information between anchors without monitoring current layers information for feature fusion

      Increased memory requirement and detection time

      Inefficient tiny face detection.

      Feature maps for small faces gradually shrink over the depth of the network and can hardly detect faces of smaller dimension.

      Region proposal is bottleneck

      The anchors are associated with last convolutional layer whose feature resolution is too weak for handling faces of diverse size.

    8. REF[40] RPN score simultaneously resulting in faster

      inferencing.

      Anchor associated layer detects faces within corresponding range of scales and has single receptive field that cannot cope up with different scales.

      TABLE VI

      Phyton packages for face detection. Details about method, weight files, configuration files, face detection function, parameters to the function and output of each method is presented. Covers both machine learning and deep learning methods.

     

    TABLE VII

    Selection of an algorithm for an application is closely related to benefits and limitations it offers. Summary related to different practical methods covering machine learning and deep learning-based approaches is presented.

    Method Benefits Limitations

    Does not work on non-frontal and occluded faces.

    Haar Cascade or Viola Jones

    HoG with SVM

    OpenCV

    CNN

    Faster detections on real time CPU

    Works on slightly rotated faces and small occlusions.

    Highly accurate than the above two methods. Higher speed attained with multiscale detection permitting resizing and optimal computations.

    Accurate and robust than conventional methods.

    Detections under varying face orientations and lighting.

    Tradeoff between false positives and minNeigbours argument. Tradeoff between inference time and scaleFactor parameter.

    Fails on profile faces and occluded faces, estimating bounding boxes debarring facial parts.

    Inference is worsened further if the input image is up sampled.

    It cannot detect faces with smaller dimension.

    High detection time limits usage for real world applications.

    Inference is worsened further if the input image is scaled using image pyramid.

    YoloFace Faster than conventional CNN. Fails for tiny faces

    MTCNN SSD

    Faster as compared to conventional convolution neural network approach using DLIB library.

    Super real time performance, this makes it unique from other methods.

    Detections are affected by extreme lightning and massive occlusions.

    Inaccurate key point regression for occluded faces.

    constraints. YOLO, MobileNet, SSD and SqueezeNet enable device computation with simple mathematics and single stage operation. Model compression with pruning, knowledge distil- lation and quantization shrink the model at the cost of reduced accuracy. An alternative to GPU is FPGA [81], allowing re- configuration leading to customized architectures and enabling model level optimization. Googles Tensor processing unit (TPU) is leveraging deep learning inference. Programming of FPGA and ASIC enforces knowledge of hardware level abstraction. OPENCL empowers software level programming of FPGA in contrast to hardware level programming.

    The potential of intelligent computing platforms for data preprocessing, model training, model deployment, parallel computing, and brisk inferencing can be harnessed in the automotive domain for a customized task. State of the art research in Computer vision for ADAS is gaining popularity and has a promising future with the integration of cameras for monitoring. The overall driver monitoring system revolves around the interpretation and processing of the video frames collected by the camera for emotion detection, drowsiness de- tection and driver authentication. Face detection is the first step towards the development of driver monitoring systems and ad- vanced driver assistance systems. Monitoring and surveillance systems require identification of passengers in the vehicle. The acquisition of frames from mltiple streams demands swift training and inference. Table IX projects machine learning based ADAS systems. Figure 12 and 13 shows computing platforms for ADAS.

  5. Challenges and Scope for Research
    1. Unconstrained Face Detection

      Unconstrained face recognition has advanced quickly, reach- ing saturation in recognition accuracy for the benchmark datasets used today. While crucial for early development, the use of a generic face detector for face imagery is a major draw- back in most benchmark datasets. Restricted variations in face posture and other confounding elements are the implication of this method. Classical face detectors toil with constraints of limited variation in pose, occlusion, expression, reflection, styling accessories. Tremendous scope towards development of exhaustive unconstrained face detection dataset and effective models is the need of an hour.

    2. Deployment on Embedded platforms

      Training of face detection model and inferencing demands fast processing on CPU or GPU or ASIC or FPGA. Some of these devices support parallel computing to enhance de- tection performance. Some edge specific tools are supported exclusively for their own computing devices. NVIDIA GPU leverages model deployment on NVIDIA specific platforms only such as NIVIDIA Jetson, NVDIA Orin, and NVIDIA Xavier. Similarly, Intel OPENVINO toolkit supports Intel chips only. Compression of the deep learning model facilitates usage on edge nodes. Parameter quantization, pruning, knowl- edge distillation, and fast exiting are popular methods for reducing the model. Although tremendous progress is ongoing for model deployment and compression, still wide scope exists for further improvement. Two avenues are open for research to

      Fig. 10. Face Detection methods

      allow heterogenous cross platform implementation on different embedded devices and compressing the model for minimal resource requirement as well as to boost accuracy and real time performance.

    3. Hardware and Software Acceleration

      To enhance the inference time, hardware suppliers are utilizing CPU, GPU, TPU, ASIC, and FPGA. CPU and GPU are flexible for diverse data handling but demands higher power consumption. In contrast, FPGA and ASIC supports reconfigurable design with lesser power requirements. FPGA programming requires hardware level knowledge and most of the researcher fail to have these fine details. OPENCL has enabled software level development of FPGA. Researchers are attracted towards software specific development due to user friendly tools, adaptable and configurable models, and command line execution interface. FPGA development can successfully enable researchers to explore for architectural variations. Data scientists can collaboratively work on hard-

      ware and software acceleration to enable configurable model design and easy implementation.

    4. Latency, Energy and Migration on Edge platforms

      Accommodation of data from many sources on cloud puts limitations on latency, energy, and migration. DNN (Deep Neural Network) partitioning performs effectively on sequen- tial design but execution for recurrent neural network (RNN) is substandard. Optimal energy consumption is key aspect for hardware platforms (GPU, TPU) and its interaction with battery management systems must be dealt with to lessen power consumption. Virtual machine and Docker containers are the solutions for migration on edge servers, handling migration of deep learning models is still an area to explore. What module of DNN model must be migrated, what modules must be retained on virtual image, and can the modules be migrated during training? Dealing with the migration, system monitoring and analysis is crucial to gain empirical knowledge of the challenges.

      TABLE VIII

      Summary for Cloud Instances with available framework, benefits, key features, customers, pricing, object storage, support for computer vision applications.

      Benefits

      Augmented AI, Autopilot, Clarify, Batch Transform, Data Wrangler, Edge Manager, Feature Store, Ground Truth, Elastic Inference, Model Monitor, Model Registry, Serverless Endpoints

      Synapse Analytics, Cognitive Services, Speech, Language, Computer Vision, Custom Vision,

       

      TABLE IX

      Machine Learning ADAS system with details about Developer, supported framework, inference engine, deployment platforms, key applications and features, support for computer vision tasks.

      Sr no About the system Developer Frameworks High-Performance

      Automotive Vision Processing and Streamlined Development Efficiency for

      Advanced Driver Assistance Systems

      Inference Engine

      DeepViewRT, TensorFlow Lite, TensorFlow

      Deployment/ Development platform

      NXP Edge Verse i.MX RT crossover MCUs,

      Key Applications

      Front View Camera, Surround View,

      Additional Applications/ Features

      Pedestrian detection,

      Lane keeping/

      departure warning

      and assistance, Traffic sign

      1. (ADAS). Accelerating

        innovation in automotive vision technology

        is fueling a transformation ADAS and will ultimately help to enable the achievement of fully autonomous L5 vehicles.

        The NVIDIA DRIVE

        platform is a full-stack solution that spans fully autonomous, highly automated, and supervised

        NXP Pytorch

        Pytorch,

        Lite

        Micro, Glow and ONNX

        Runtime

        and i.MX family application processors

        NVIDIA DRIVE

        Hyperion,

        Driver monitoring

        system,

        Driver Occupant system

        Autonomous</>

        recognition (TSR),

        Collision warning and avoidance, Blind spot monitoring

        Active safety, Highway driving,

      2. driving. It includes active

        safety, automated driving,

        and parking-plus AI cockpit capabilities-scaling from Level 2+ to the highest levels of autonomy for large-scale production.

        More customers, across a diverse set of industries, choose AWS compared to any other cloud to build,

        train, and deploy their

        machine learning (ML) applications. AWS

      3. delivers the broadest

      choice of powerful compute, high speed networking, and scalable high performance storage options for any ML project or application.

       

      Fig. 11. Inefficient face detection affected by network architecture and performance metrics.

      Fig. 12. NVIDIA GPU accelerated Deep Learning frameworks.

    5. Interpretable Deep Learning models

      Deep learning has set a performance benchmark on chal- lenging face detection datasets, there many question that are unanswered. What is the underlying mechanism for the model? What is the relation between number of parameters and accu- racy? How to attain maximum accuracy with minimal change in hyperparameters? What is effect of number of epochs and loss function in attaining a global minimum? We can monitor some of these features during training by using TensorFlow logs, yet a concise understanding about the interpretation of different components will yield a better face detection model.

  6. Conclusion

We have surveyed different modules of end-to-end real world deep learning application development impinged on

Fig. 13. NXP Software Development for Machine Learning

diverse face detection architectures which have gained pop- ularity on various benchmarking datasets. We have discussed the evolution stages for face detections methods and the critical improvements over predecessor ones. We have partitioned these models in various categories depending on architecture such as CNN, SSD, Yolo, and others, and have reported their major contributions for boosting the detection performance. An overview of practical software packages highlighting the package size, benefits and, drawbacks enabling the developer to curate appropriate algorithm for their application. We have summarized the need to utilize cloud instances and machine learning ADAS development tools due emergence of computer vision systems in driver monitoring systems from security aspect. Further, we have presented the challenges and oppor-

tunities for face detection systems.

ACKNOWLEDGMENT

The authors express sincere gratitude to Center of Excel- lence in Signal and Image Processing, COEP Tech. Univer- sity, Pune, India and Fourfront Private Ltd, Pune, India for providing the facilities and funding for conducting the survey.

REFERENCES

  1. C. Machinery, Computing machinery and intelligence-am turing,

    Mind, vol. 59, no. 236, p. 433, 1950.

  2. M. C. Burl, T. K. Leung, and P. Perona, Face localization via shape statistics, in Internatational Workshop on Automatic Face and Gesture Recognition. University of Zurich, 1995, pp. 154159.
  3. S.-H. Jeng, H. Y. M. Liao, C. C. Han, M. Y. Chern, and Y. T. Liu, Facial feature detection using geometrical face model: An efficient approach, Pattern recognition, vol. 31, no. 3, pp. 273282, 1998.
  4. S. Tsekeridou and I. Pitas, Facial feature extraction in frontal views us- ing biometric analogies, in 9th European Signal Processing Conference (EUSIPCO 1998). IEEE, 1998, pp. 14.
  5. K. C. Yow and R. Cipolla, Feature-based human face detection, Image and vision computing, vol. 15, no. 9, pp. 713735, 1997.
  6. C. Garcia and G. Tziritas, Face detection using quantized skin color regions merging and wavelet packet analysis, IEEE Transactions on multimedia, vol. 1, no. 3, pp. 264277, 1999.
  7. D. G. Lowe, Object recognition from local scale-invariant features, in Proceedings of the seventh IEEE international conference on computer vision, vol. 2. Ieee, 1999, pp. 11501157.
  8. H. A. Rowley, S. Baluja, and T. Kanade, Neural network-based face detection, IEEE Transactions on pattern analysis and machine intelligence, vol. 20, no. 1, pp. 2338, 1998.
  9. C. Garcia and M. Delakis, Convolutional face finder: A neural archi- tecture for fast and robust face detection, IEEE Transactions on pattern analysis and machine intelligence, vol. 26, no. 11, pp. 14081423, 2004.
  10. P. Viola and M. Jones, Rapid object detection using a boosted cascade of simple features, in Proceedings of the 2001 IEEE computer society conference on computer vision and pattern recognition. CVPR 2001, vol.
    1. Ieee, 2001, pp. II.
  11. Y.-Q. Wang, An analysis of the viola-jones face detection algorithm,

    Image Processing On Line, vol. 4, pp. 128148, 2014.

  12. H. Bay, T. Tuytelaars, and L. Van Gool, Surf: Speeded up robust features, Lecture notes in computer science, vol. 3951, pp. 404417, 2006.
  13. K. Vikram and S. Padmavathi, Facial parts detection using viola jones algorithm, in 2017 4th international conference on advanced computing and communication systems (ICACCS). IEEE, 2017, pp. 14.
  14. L. Cuimei, Q. Zhiliang, J. Nan, and W. Jianhua, Human face detection algorithm via haar cascade classifier combined with three additional classifiers, in 2017 13th IEEE International Conference on Electronic Measurement & Instruments (ICEMI). IEEE, 2017, pp. 483487.
  15. J. Huang, Y. Shang, and H. Chen, Improved viola-jones face detection algorithm based on hololens, EURASIP Journal on Image and Video Processing, vol. 2019, no. 1, pp. 111, 2019.
  16. N. Dalal and B. Triggs, Histograms of oriented gradients for human detection, in 2005 IEEE computer society conference on computer vision and pattern recognition (CVPR05), vol. 1. Ieee, 2005, pp. 886893.
  17. N. Dalal, B. Triggs, and C. Schmid, Human detection using oriented histograms of flow and appearance, in Computer VisionECCV 2006: 9th European Conference on Computer Vision, Graz, Austria, May 7-13, 2006. Proceedings, Part II 9. Springer, 2006, pp. 428441.
  18. Q. Zhu, M.-C. Yeh, K.-T. Cheng, and S. Avidan, Fast human detection using a cascade of histograms of oriented gradients, in 2006 IEEE computer society conference on computer vision and pattern recognition (CVPR06), vol. 2. IEEE, 2006, pp. 14911498.
  19. D. E. King, Max-margin object detection, arXiv preprint arXiv:1502.00046, 2015.
  20. X. Wang, T. X. Han, and S. Yan, An hog-lbp human detector with partial occlusion handling, in 2009 IEEE 12th international conference on computer vision. IEEE, 2009, pp. 3239.
  21. W. Liu, D. Anguelov, D. Erhan, C. Szegedy, S. Reed, C.-Y. Fu, and

    p>A. C. Berg, Ssd: Single shot multibox detector, in Computer Vision ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, October 1114, 2016, Proceedings, Part I 14. Springer, 2016, pp.

    2137.

  22. S. Zhang, X. Zhu, Z. Lei, H. Shi, X. Wang, and S. Z. Li, S3fd: Single shot scale-invariant face detector, in Proceedings of the IEEE international conference on computer vision, 2017, pp. 192201.
  23. J. Zhang, X. Wu, S. C. Hoi, and J. Zhu, Feature agglomeration networks for single stage face detection, Neurocomputing, vol. 380, pp. 180189, 2020.
  24. M. Najibi, P. Samangouei, R. Chellappa, and L. S. Davis, Ssh: Single stage headless face detector, in Proceedings of the IEEE international conference on computer vision, 2017, pp. 48754884.
  25. J. Li, Y. Wang, C. Wang, Y. Tai, J. Qian, J. Yang, C. Wang, J. Li, and
    1. Huang, Dsfd: dual shot face detector, in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2019, pp. 5060 5069.
  26. R. J. Wang, X. Li, and C. X. Ling, Pelee: A real-time object detection system on mobile devices, Advances in neural information processing systems, vol. 31, 2018.
  27. X. Tang, D. K. Du, Z. He, and J. Liu, Pyramidbox: A context-assisted single shot face detector, in Proceedings of the European conference on computer vision (ECCV), 2018, pp. 797813.
  28. S. Zhang, L. Wen, H. Shi, Z. Lei, S. Lyu, and S. Z. Li, Single-shot scale- aware network for real-time face detection, International Journal of Computer Vision, vol. 127, pp. 537559, 2019.
  29. V. Bazarevsky, Y. Kartynnik, A. Vakunov, K. Raveendran, and

    M. Grundmann, Blazeface: Sub-millisecond neural face detection on mobile gpus, arXiv preprint arXiv:1907.05047, 2019.

  30. Z. Tian, C. Shen, H. Chen, and T. He, Fcos: Fully convolutional one- stage object detection, in Proceedings of the IEEE/CVF international conference on computer vision, 2019, pp. 96279636.
  31. Y. He, D. Xu, L. Wu, M. Jian, S. Xiang, and C. Pan, Lffd: A light and fast face detector for edge devices, arXiv preprint arXiv:1904.10633, 2019.
  32. Z. Li, X. Tang, J. Han, J. Liu, and R. He, Pyramidbox++: High perfor- mance detector for finding tiny face, arXiv preprint arXiv:1904.00386, 2019.
  33. S. Liu, D. Huang et al., Receptive field block net for accurate and fast object detection, in Proceedings of the European conference on computer vision (ECCV), 2018, pp. 385400.
  34. L. Shi, X. Xu, and I. A. Kakadiaris, Sanet: Smoothed attention network for single stage face detector, in 2019 International Conference on Biometrics (ICB). IEEE, 2019, pp. 17.
  35. R. Girshick, J. Donahue, T. Darrell, and J. Malik, Rich feature hierarchies for accurate object detection and semantic segmentation, in Proceedings of the IEEE conference on computer vision and pattern recognition, 2014, pp. 580587.
  36. H. Li, Z. Lin, X. Shen, J. Brandt, and G. Hua, A convolutional neural network cascade for face detection, in Proceedings of the IEEE conference on computer vision and pattern recognition, 2015, pp. 5325 5334.
  37. H. Jiang and E. Learned-Miller, Face detection with the faster r-cnn, in 2017 12th IEEE international conference on automatic face & gesture recognition (FG 2017). IEEE, 2017, pp. 650657.
  38. S. Ren, K. He, R. Girshick, and J. Sun, Faster r-cnn: Towards real-time object detection with region proposal networks, Advances in neural information processing systems, vol. 28, 2015.
  39. S. Yang, P. Luo, C.-C. Loy, and X. Tang, From facial parts responses to face detection: A deep learning approach, in Proceedings of the IEEE international conference on computer vision, 2015, pp. 36763684.
  40. S. Yang, P. Luo, C. C. Loy, and X. Tang, Faceness-net: Face detection through deep facial part responses, IEEE transactions on pattern analysis and machine intelligence, vol. 40, no. 8, pp. 18451859, 2017.
  41. C. Zhu, Y. Zheng, K. Luu, and M. Savvides, Cms-rcnn: contextual multi- scale region-based cnn for unconstrained face detection, Deep learning for biometrics, pp. 5779, 2017.
  42. R. Ranjan, V. M. Patel, and R. Chellappa, Hyperface: A deep multi- task learning framework for face detection, landmark localization, pose estimation, and gender recognition, IEEE transactions on pattern analysis and machine intelligence, vol. 41, no. 1, pp. 121135, 2017.
  43. D. Chen, G. Hua, F. Wen, and J. Sun, Supervised transformer network for efficient face detection, in Computer VisionECCV 2016: 14th

    European Conference, Amsterdam, The Netherlands, October 11-14, 2016, Proceedings, Part V 14. Springer, 2016, pp. 122138.

  44. Y. Wang, X. Ji, Z. Zhou, H. Wang, and Z. Li, Detecting faces using region-based fully convolutional networks, arXiv preprint arXiv:1709.05256, 2017.
  45. K. Zhang, Z. Zhang, H. Wang, Z. Li, Y. Qiao, and W. Liu, Detecting faces using inside cascaded contextual cnn, in Proceedings of the IEEE international conference on computer vision, 2017, pp. 31713179.
  46. S. Yang, Y. Xiong, C. C. Loy, and X. Tang, Face detection through scale-friendly deep convolutional networks, arXiv preprint arXiv:1706.02863, 2017.
  47. C. Zhang, X. Xu, and D. Tu, Face detection using improved faster rcnn,

    arXiv preprint arXiv:1802.02142, 2018.

  48. H. Wang, Z. Li, X. Ji, and Y. Wang, Face r-cnn, arXiv preprint arXiv:1706.01061, 2017.
  49. S. Zhang, X. Zhu, Z. Lei, H. Shi, X. Wang, and S. Z. Li, Faceboxes: A cpu real-time face detector with high accuracy, in 2017 IEEE International Joint Conference on Biometrics (IJCB). IEEE, 2017,

    pp. 19.

  50. W. Wu, Y. Yin, X. Wang, and D. Xu, Face detection with different scales based on faster r-cnn, IEEE transactions on cybernetics, vol. 49, no. 11,

    pp. 40174028, 2018.

  51. P. Jiang, D. Ergu, F. Liu, Y. Cai, and B. Ma, A review of yolo algorithm developments, Procedia Computer Science, vol. 199, pp. 10661073, 2022.
  52. D. Garg, P. Goel, S. Pandya, A. Ganatra, and K. Kotecha, A deep learning approach for face detection using yolo, in 2018 IEEE Punecon. IEEE, 2018, pp. 14.
  53. W. Yang and Z. Jiachun, Real-time face detection based on yolo, in 2018 1st IEEE international conference on knowledge innovation and invention (ICKII). IEEE, 2018, pp. 221224.
  54. D. Qi, W. Tan, Q. Yao, and J. Liu, Yolo5face: Why reinventing a face detector, in European Conference on Computer Vision. Springer, 2022,

    pp. 228244.

  55. W. Chen, H. Huang, S. Peng, C. Zhou, and C. Zhang, Yolo-face: a real- time face detector, The Visual Computer, vol. 37, pp. 805813, 2021.
  56. D. T. Nguyen, T. N. Nguyen, H. Kim, and H.-J. Lee, A high- throughput and power-efficient fpga implementation of yolo cnn for object detection, IEEE Transactions on Very Large Scale Integration (VLSI) Systems, vol. 27, no. 8, pp. 18611873, 2019.
  57. S. Zhang, J. Cao, Q. Zhang, Q. Zhang, Y. Zhang, and Y. Wang, An fpga- based reconfigurable cnn accelerator for yolo, in 200 IEEE 3rd International Conference on Electronics Technology (ICET). IEEE, 2020,

    pp. 7478.

  58. R. Huang, J. Pedoeem, and C. Chen, Yolo-lite: a real-time object detection algorithm optimized for non-gpu computers, in 2018 IEEE international conference on big data (big data). IEEE, 2018, pp. 2503 2510.
  59. Y. Zhu, H. Cai, S. Zhang, C. Wang, and Y. Xiong, Tinaface: Strong but simple baseline for face detection, arXiv preprint arXiv:2011.13183, 2020.
  60. J. Deng, J. Guo, E. Ververas, I. Kotsia, and S. Zafeiriou, Retinaface: Single-shot multi-level face localisation in the wild, in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2020,

    pp. 52035212.

  61. K. Lin, H. Zhao, J. Lv, C. Li, X. Liu, R. Chen, and R. Zhao, Face detection and segmentation based on improved mask r-cnn, Discrete dynamics in nature and society, vol. 2020, pp. 111, 2020.
  62. S. Zhang, C. Chi, Z. Lei, and S. Z. Li, Refineface: Refinement neural network for high performance face detection, IEEE transactions on pattern analysis and machine intelligence, vol. 43, no. 11, pp. 4008 4020,

    2020.

  63. K. Liu, H. Tang, S. He, Q. Yu, Y. Xiong, and N. Wang, Performance validation of yolo variants for object detection, in Proceedings of the 2021 International Conference on bioinformatics and intelligent computing, 2021, pp. 239243.
  64. J. Redmon and A. Farhadi, Yolo9000: better, faster, stronger, in Proceedings of the IEEE conference on computer vision and pattern recognition, 2017, pp. 72637271.
  65. P. Adarsh, P. Rathi, and M. Kumar, Yolo v3-tiny: Object detection and recognition using one stage improved model, in 2020 6th interna- tional conference on advanced computing and communication systems (ICACCS). IEEE, 2020, pp. 687694.
  66. A. Bochkovskiy, C.-Y. Wang, and H.-Y. M. Liao, Yolov4: Op- timal speed and accuracy of object detection, arXiv preprint arXiv:2004.10934, 2020.
  67. C. Gupta, N. S. Gill, P. Gulia, and J. M. Chatterjee, A novel finetuned yolov6 transfer learning model for real-time object detection, Journal of Real-Time Image Processing, vol. 20, no. 3, p. 42, 2023.
  68. C.-Y. Wang, A. Bochkovskiy, and H.-Y. M. Liao, Yolov7: Trainable bag- of-freebies sets new state-of-the-art for real-time object detectors, in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2023, pp. 74647475.
  69. M. Khan, S. Chakraborty, R. Astya, and S. Khepra, Face detection and recognition using opencv, in 2019 International Conference on Computing, Communication, and Intelligent Systems (ICCCIS). IEEE, 2019, pp. 116 119.
  70. C. Zhang and Z. Zhang, Improving multiview face detection with multi- task deep convolutional neural networks, in IEEE Winter Conference on Applications of Computer Vision. IEEE, 2014, pp. 10361041.
  71. K. Zhang, Z. Zhang, Z. Li, and Y. Qiao, Joint face detection and alignment using multitask cascaded convolutional networks, IEEE signal processing letters, vol. 23, no. 10, pp. 14991503, 2016.
  72. J. Xiang and G. Zhu, Joint face detection and facial expression recog- nition with mtcnn, in 2017 4th international conference on information science and control engineering (ICISCE). IEEE, 2017, pp. 424427.
  73. P. Hu and D. Ramanan, Finding tiny faces, in Proceedings of the IEEE conference on computer vision and pattern recognition, 2017, pp. 951959.
  74. C. Zhu, R. Tao, K. Luu, and M. Savvides, Seeing small faces from robust anchors perspective, in Proceedings of the IEEE conference on computer vision and pattern recognition, 2018, pp. 51275136.
  75. V. Albiero, X. Chen, X. Yin, G. Pang, and T. Hassner, img2pose: Face alignment and detection via 6dof, face pose estimation, in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2021, pp. 76177627.
  76. T. Baltrusaitis, P. Robinson, and L.-P. Morency, Openface: an open source facial behavior analysis toolkit, in 2016 IEEE winter conference on applications of computer vision (WACV). IEEE, 2016, pp. 110.
  77. G. Ghiasi and C. C. Fowlkes, Occlusion coherence: Detecting and localizing occluded faces, arXiv preprint arXiv:1506.08347, 2015.
  78. X. Shi, S. Shan, M. Kan, S. Wu, and X. Chen, Real-time rotation- invariant face detection with progressive calibration networks, in Pro- ceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2018, pp. 22952303.
  79. J. Chen and X. Ran, Deep learning with edge computing: A review,

    Proceedings of the IEEE, vol. 107, no. 8, pp. 16551674, 2019.

  80. H. Hua, Y. Li, T. Wang, N. Dong, W. Li, and J. Cao, Edge comput- ing with artificial intelligence: A machine learning perspective, ACM Computing Surveys, vol. 55, no. 9, pp. 135, 2023.
  81. G. Lacey, G. W. Taylor, and S. Areibi, Deep learning on fpgas: Past, present, and future, arXiv preprint arXiv:1602.04283, 2016.