Abstract

We incorporate deep learning techniques into capacitive images of body parts (ear, four fingers, and thumb) to improve the performance of user authentication in smartphones. Use of a capacitive touchscreen as an image sensor has several advantages, such as it is less sensitive to poor illumination conditions, occlusions, and pose variations. Also, it does not need an additional hardware like iris or fingerprint scanner. Use of capacitive images for user authentication is not new. However, the performance, specially, false reject rates (FRRs) of the state-of-the-art capacitive image-based systems are poor. In this paper, we focus on improving the performance and leverage deep learning. Deep learning techniques demonstrated spectacular performance in previous physical biometrics-based research. However, to our knowledge, effectiveness of deep learning is still unexplored in capacitive touchscreen-based user authentication. In order to bridge this research gap, we devise a multi-modal deep learning model, namely UASNet, and compare its performance with a large set of uni- and multi-modal baselines. Using the UASNet, we achieve an accuracy of 99.77%, an EER of 0.48%, and an FRR of 1.19% at FAR of 0.06%.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call