3260 papers • 126 benchmarks • 313 datasets
Face identification is the task of matching a given face image to one in an existing database of faces. It is the second part of face recognition (the first part being detection). It is a one-to-many mapping: you have to find an unknown person in a database to find who that person is.
(Image credit: Papersgraph)
These leaderboards are used to track progress in face-identification
Use these libraries to find face-identification models and implementations
No subtasks available.
A system that directly learns a mapping from face images to a compact Euclidean space where distances directly correspond to a measure offace similarity, and achieves state-of-the-art face recognition performance using only 128-bytes perface.
One of the main challenges in feature learning using Deep Convolutional Neural Networks (DCNNs) for large-scale face recognition is the design of appropriate loss functions that can enhance the discriminative power. Centre loss penalises the distance between deep features and their corresponding class centres in the Euclidean space to achieve intra-class compactness. SphereFace assumes that the linear transformation matrix in the last fully connected layer can be used as a representation of the class centres in the angular space and therefore penalises the angles between deep features and their corresponding weights in a multiplicative way. Recently, a popular line of research is to incorporate margins in well-established loss functions in order to maximise face class separability. In this paper, we propose an Additive Angular Margin Loss (ArcFace) to obtain highly discriminative features for face recognition. The proposed ArcFace has a clear geometric interpretation due to its exact correspondence to geodesic distance on a hypersphere. We present arguably the most extensive experimental evaluation against all recent state-of-the-art face recognition methods on ten face recognition benchmarks which includes a new large-scale image database with trillions of pairs and a large-scale video dataset. We show that ArcFace consistently outperforms the state of the art and can be easily implemented with negligible computational overhead. To facilitate future research, the code has been made available.
This paper proposes the angular softmax (A-Softmax) loss that enables convolutional neural networks (CNNs) to learn angularly discriminative features in deep face recognition (FR) problem under open-set protocol.
Experimental results show that the proposed framework can utilize large-scale noisy data to learn a Light model that is efficient in computational costs and storage spaces and achieves state-of-the-art results on various face benchmarks without fine-tuning.
With enhanced local modeling via the micro network, the proposed deep network structure NIN is able to utilize global average pooling over feature maps in the classification layer, which is easier to interpret and less prone to overfitting than traditional fully connected layers.
This paper reformulates the softmax loss as a cosine loss by L2 normalizing both features and weight vectors to remove radial variations, based on which acosine margin term is introduced to further maximize the decision margin in the angular space, and achieves minimum intra-class variance and maximum inter- class variance by virtue of normalization and cosine decision margin maximization.
This paper proposes two very deep neural network architectures, referred to as DeepID3, for face recognition, rebuilt from stacked convolution and inception layers proposed in VGG net and GoogLeNet to make them suitable to face recognition.
This work theoretically analyze the upper limit of model parallelism in face recognition and proposes a load-balanced sparse distributed classification training method, Partial FC, which is capable of using a machine with only 8 Nvidia Tesla V100 GPUs to implement training on a face recognition data set with up to 29 million IDs.
The FacePoseNet (FPN) is claimed to be a far faster and far more accurate face alignment method than using facial landmark detectors, and aligns faces at a small fraction of the computational cost of comparably accurate landmark detectors.
Evaluation of GhostFaceNets using various benchmarks reveals that these models offer superior performance while requiring a computational complexity significantly lower than that of State-Of-The-Art (SOTA) big convolutional neural network (CNN) models, which can require hundreds of millions of FLOPs.
Adding a benchmark result helps the community track progress.