Face detection generally requires prior boxes and an extra non-maximum suppression(NMS) post-processing in modern deep learning methods. However, anchor design and anchor matching strategy ...significantly affect the performance of face detectors, so we have to spend a lot of time on anchor designing for different business scenarios. The other issue is that NMS cannot be easily parallelized and it may become a bottleneck of detection speed. In this paper, we propose a simple yet efficient pure convolutional neural network face detection method, named dual-branch center face detector(DBCFace for short), which solve face detection via a dual branch fully convolutional framework without extra anchor design and NMS. Extensive experiments are conducted on four popular face detection benchmarks, including AFW, PASCAL face, FDDB, and WIDER FACE, demonstrating that our method is comparable with state-of-the-art methods while the speed is faster.
Current face or object detection methods via convolutional neural network (such as OverFeat, R-CNN and DenseNet) explicitly extract multi-scale features based on an image pyramid. However, such a ...strategy increases the computational burden for face detection. In this paper, we propose a fast face detection method based on discriminative complete features (DCFs) extracted by an elaborately designed convolutional neural network, where face detection is directly performed on the complete feature maps. DCFs have shown the ability of scale invariance, which is beneficial for face detection with high speed and promising performance. Therefore, extracting multi-scale features on an image pyramid employed in the conventional methods is not required in the proposed method, which can greatly improve its efficiency for face detection. Experimental results on several popular face detection datasets show the efficiency and the effectiveness of the proposed method for face detection.
In recent years, face detection has achieved considerable attention in the field of computer vision using traditional machine learning techniques and deep learning techniques. Deep learning is used ...to build the most recent and powerful face detection algorithms. However, partial face detection still remains to achieve remarkable performance. Partial faces are occluded due to hair, hat, glasses, hands, mobile phones, and side-angle-captured images. Fewer facial features can be identified from such images. In this paper, we present a deep convolutional neural network face detection method using the anchor boxes section strategy. We limited the number of anchor boxes and scales and chose only relevant to the face shape. The proposed model was trained and tested on a popular and challenging face detection benchmark dataset, i.e., Face Detection Dataset and Benchmark (FDDB), and can also detect partially covered faces with better accuracy and precision. Extensive experiments were performed, with evaluation metrics including accuracy, precision, recall, F1 score, inference time, and FPS. The results show that the proposed model is able to detect the face in the image, including occluded features, more precisely than other state-of-the-art approaches, achieving 94.8% accuracy and 98.7% precision on the FDDB dataset at 21 frames per second (FPS).
In recent years, generative adversarial networks (GANs) have been widely used to generate realistic fake face images, which can easily deceive human beings. To detect these images, some methods have ...been proposed. However, their detection performance will be degraded greatly when the testing samples are post-processed. In this paper, some experimental studies on detecting post-processed GAN-generated face images find that (a) both the luminance component and chrominance components play an important role, and (b) the RGB and YCbCr color spaces achieve better performance than the HSV and Lab color spaces. Therefore, to enhance the robustness, both the luminance component and chrominance components of dual-color spaces (RGB and YCbCr) are considered to utilize color information effectively. In addition, the convolutional block attention module and multilayer feature aggregation module are introduced into the Xception model to enhance its feature representation power and aggregate multilayer features, respectively. Finally, a robust dual-stream network is designed by integrating dual-color spaces RGB and YCbCr and using an improved Xception model. Experimental results demonstrate that our method outperforms some existing methods, especially in its robustness against different types of post-processing operations, such as JPEG compression, Gaussian blurring, gamma correction, and median filtering.
In this article, we propose a model of face detection in risk situations to help rescue teams speed up the search of people who might need help. The proposed lightweight convolutional neural network ...(CNN) architecture is designed to detect faces of people in mines, avalanches, under water, or other dangerous situations when their face might not be very visible over surrounding background. We have designed a novel light architecture cooperating with the proposed sliding window procedure. The designed model works with maximum simplicity to support mobile devices. An output from processing presents a box on face location in the screen of device. The model was trained by using Adam and tested on various images. Results show that proposed lightweight CNN detects human faces over various textures with accuracy above 99% and precision above 98% what proves the efficiency of our proposed model.
Abstract
Recognition of facial expression has many potential applications that have attracted the researcher’s attention during the last decade. Taking out of features, is an important step in the ...analysis of expression that contributes to a quick and accurate recognition of expression, i.e., happiness, surprise and disgust, sadness, anger, and fear are expressions of the faces. Facial expressions are most frequently used to interpret human emotions. Two categories contain a range of different emotions: positive emotions and non-positive emotions. Face Detection, Extraction, Classification, and Recognition are major steps used in the proposed system. The proposed segmentation techniques are applied and compared to determine which method is appropriate for splitting the mouth region, and then the mouth region can be extracted using techniques for stretching contrasts and segmenting the image. After the extraction of the mouth area, the facial emotions are graded in the face picture region of the extracted mouth based on white pixel values. The Supervisory Learning Approach is widely used for face identification algorithms and it takes more computation time and effort. It may also give incorrect class labels in the classification process. For this reason, supervised learning and reinforcement learning is being used. In general, it will be like a trial-and-error method that is, in the training process it tries to learn and produce expected results. It was specified accordingly. Reinforcement learning always tries to enhance the results.
Student engagement is a key element to ensure effective learning process. In this work, we presented an automatic system for monitoring engagement level from students’ facial gestures. In this way, ...the tutor can analyse the engagement level of students and improve the teaching method and strategies to enhance learning process. There has been extensive research on automated classification of engagement level, but most of these methods rely mainly on expensive eye trackers or physiological sensors in controlled settings. The proposed system monitors and classifies engagement level of student based on YOLO algorithm by determining facial gestures, where students move freely and respond naturally to lectures and surroundings. The proposed model gives a mean average precision (mAP) of 0.65 on a complex dataset where students were allowed to move freely during lecture.
Generative Adversarial Network (GAN) based techniques can generate and synthesize realistic faces that cause profound social concerns and security problems. Existing methods for detecting ...GAN-generated faces can perform well on limited public datasets. However, images from existing datasets do not represent real-world scenarios well enough in terms of view variations and data distributions, where real faces largely outnumber synthetic ones. The state-of-the-art methods do not generalize well in real-world problems and lack the interpretability of detection results. Performance of existing GAN-face detection models degrades accordingly when facing data imbalance issues. To address these shortcomings, we propose a robust, attentive, end-to-end framework that spots GAN-generated faces by analyzing eye inconsistencies. Our model automatically learns to identify inconsistent eye components by localizing and comparing artifacts between eyes. After the iris regions are extracted by Mask-RCNN, we design a Residual Attention Network (RAN) to examine the consistency between the corneal specular highlights of the two eyes. Our method can effectively learn from imbalanced data using a joint loss function combining the traditional cross-entropy loss with a relaxation of the ROC-AUC loss via Wilcoxon-Mann-Whitney (WMW) statistics. Comprehensive evaluations on a newly created FFHQ-GAN dataset in both balanced and imbalanced scenarios demonstrate the superiority of our method.
The Global Novel Coronavirus Disease-2019 (COVID-19) pandemic has forced social distancing norms that have been followed worldwide. Thus, traditional biometric-based attendance marking systems are ...replaced with contactless attendance marking schemes. However, there are limitations of manufacturing cost, spoofing attacks, and security vulnerabilities. Thus, the paper proposes a contactless camera-based attendance system with the equipped functionalities of anti-spoofing. The proposed scheme can detect liveliness, so fake attendance marking is eliminated. The proposed scheme is also scalable and cost-effective, with generic solutions adaptable to schools, colleges, or other places where attendance is required. The system also eliminates the limitation of one-entry by multiple face-marking systems that allow simultaneous attendance marking. In performance analysis, parameters like image precision, storage cost, retrieval latency, and analysis of the anti-spoofing module is presented against existing schemes. An accuracy of 95.85% is reported for the model, with a significant improvement of 33.52% in storage cost through the Firebase database, which outperforms existing state-of-the-art schemes.
•We design the Rapidly Digested Convolution Layers (RDCL) to enable face detection to achieve CPU real-time speed.•We introduce the Multiple Scale Convolution Layers (MSCL) to handle various scales ...of face via enriching features and discretizing anchors over layers.•We propose a new anchor densification strategy to improve the recall rate of small faces.•We present a Divide and Conquer Head (DCH) to boost the prediction ability of the detection layer.•We achieve state-of-the-art performance on the AFW, PASCAL face, FDDB and WIDER FACE datasets among CPU real-time methods.
Although tremendous strides have been made in face detection, one of the remaining open issues is to achieve CPU real-time speed as well as maintain high performance, since effective models for face detection tend to be computationally prohibitive. To address this issue, we propose a novel face detector, named FaceBoxes, with superior performance on both speed and accuracy. Specifically, the proposed method has a lightweight yet powerful network that consists of the Rapidly Digested Convolution Layers (RDCL) and the Multiple Scale Convolution Layers (MSCL). The former is designed to enable FaceBoxes to achieve CPU real-time speed, while the latter aims to enrich the features and discretize anchors over different layers to handle faces of various scales. Besides, we propose a new anchor densification strategy to make different types of anchors have the same density on the image, which significantly improves the recall rate of small faces. Finally, we present a Divide and Conquer Head (DCH) to boost the prediction ability of the detection layer using above strategy. As a consequence, the proposed detector runs at 28 FPS on the CPU and 254 FPS using a GPU for VGA-resolution images. Moreover, the speed of FaceBoxes is invariant to the number of faces. We evaluate the proposed method on several face detection benchmarks including AFW, PASCAL face, FDDB, WIDER FACE and achieve state-of-the-art performance among CPU real-time methods.