Hostname: page-component-77f85d65b8-6bnxx Total loading time: 0 Render date: 2026-03-27T09:36:39.644Z Has data issue: false hasContentIssue false

The future of biometrics technology: from face recognition to related applications

Published online by Cambridge University Press:  28 May 2021

Hitoshi Imaoka*
Affiliation:
NEC Corporation, Minato-ku, Tokyo, Japan
Hiroshi Hashimoto
Affiliation:
NEC Corporation, Minato-ku, Tokyo, Japan
Koichi Takahashi
Affiliation:
NEC Corporation, Minato-ku, Tokyo, Japan
Akinori F. Ebihara
Affiliation:
NEC Corporation, Minato-ku, Tokyo, Japan
Jianquan Liu
Affiliation:
NEC Corporation, Minato-ku, Tokyo, Japan
Akihiro Hayasaka
Affiliation:
NEC Corporation, Minato-ku, Tokyo, Japan
Yusuke Morishita
Affiliation:
NEC Corporation, Minato-ku, Tokyo, Japan
Kazuyuki Sakurai
Affiliation:
NEC Corporation, Minato-ku, Tokyo, Japan
*
Corresponding author: Hitoshi Imaoka Email: h-imaoka_cb@nec.com

Abstract

Biometric recognition technologies have become more important in the modern society due to their convenience with the recent informatization and the dissemination of network services. Among such technologies, face recognition is one of the most convenient and practical because it enables authentication from a distance without requiring any authentication operations manually. As far as we know, face recognition is susceptible to the changes in the appearance of faces due to aging, the surrounding lighting, and posture. There were a number of technical challenges that need to be resolved. Recently, remarkable progress has been made thanks to the advent of deep learning methods. In this position paper, we provide an overview of face recognition technology and introduce its related applications, including face presentation attack detection, gaze estimation, person re-identification and image data mining. We also discuss the research challenges that still need to be addressed and resolved.

Information

Type
Industrial Technology Advances
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted re-use, distribution, and reproduction in any medium, provided the original work is properly cited.
Copyright
Copyright © The Author(s), 2021 published by Cambridge University Press in association with Asia Pacific Signal and Information Processing Association
Figure 0

Fig. 1. Face recognition processing.

Figure 1

Fig. 2. Processing flow of the face detection proposed in 2005 [5].

Figure 2

Fig. 3. Processing flow of the face/human body detection technology based on a deep neural network.

Figure 3

Fig. 4. Face alignment for detecting the feature points of facial parts.

Figure 4

Fig. 5. Feature extraction with a CNN for the face matching.

Figure 5

Fig. 6. Example of presentation attack types. (a) 2D print attack. (b) 2D replay attack. (c) 3D spoofing mask attack.

Figure 6

Fig. 7. Speed and accuracy test results, adapted from [42]. (a) Two-way ANOVA comparing SpecDiff and ResNet4. (s)BPCER and (s)ACER indicate that sBPCER and sACER are used for evaluating public databases, while genuine BPCER and ACER are used for evaluating the in-house database. Resulting p-values show statistical significance in APCER and (s)ACER. (b) Summary of execution speeds. The proposed SpecDiff descriptor classified with the SVM RBF kernel is compared with ResNet4. Execution speeds on iPhone7, iPhone XR, and iPad Pro are measured.

Figure 7

Table 1. Mean validation errors of selected algorithms.

Figure 8

Fig. 8. Average detection error tradeoff (DET) curves across 10-fold cross-validation trials, adapted from [42]. Implicit3D is a different flash-based algorithm [46] included for comparison. (1) NUAA database. (2) Replay-Attack database. (3) SiW database.

Figure 9

Fig. 9. Outline of remote gaze estimation technology.

Figure 10

Fig. 10. Examples of non-mate image pairs with similar backgrounds. Images are selected from VIPeR [60] dataset.

Figure 11

Fig. 11. Workflow of dropout technique. In the first stage, a color image is input and the saliency map is calculated. This map is used in the next stage along with the original image to learn robust features by a CNN. The output is shown as a vector of IDs, but the CNN codes learned in the penultimate layer can be used as the extracted features.

Figure 12

Table 2. CMC accuracies on VIPeR [60].

Figure 13

Fig. 12. An example of face grouping by Luigi index [64].

Figure 14

Fig. 13. Visualization results [65] of loiterer candidates discovered by AntiLoiter [63].

Figure 15

Fig. 14. Examples of behavior patterns regarding potential loiterers.

Figure 16

Fig. 15. System VisLoiter+ [68] implemented with the models proposed in [67].

Figure 17

Fig. 16. An example of stalking scenario [69].