Faseela Chakkalakkal Abdullakutty
Unmasking the imposters: towards improving the generalisation of deep learning methods for face presentation attack detection.
Abdullakutty, Faseela Chakkalakkal
Abstract
Identity theft has had a detrimental impact on the reliability of face recognition, which has been extensively employed in security applications. The most prevalent are presentation attacks. By using a photo, video, or mask of an authorized user, attackers can bypass face recognition systems. Fake presentation attacks are detected by the camera sensors of face recognition systems using face presentation attack detection. Presentation attacks can be detected using convolutional neural networks, commonly used in computer vision applications. An in-depth analysis of current deep learning methods is used in this research to examine various aspects of detecting face presentation attacks. A number of new techniques are implemented and evaluated in this study, including pre-trained models, manual feature extraction, and data aggregation. The thesis explores the effectiveness of various machine learning and deep learning models in improving detection performance by using publicly available datasets with different dataset partitions than those specified in the official dataset protocol. Furthermore, the research investigates how deep models and data aggregation can be used to detect face presentation attacks, as well as a novel approach that combines manual features with deep features in order to improve detection accuracy. Moreover, task-specific features are also extracted using pre-trained deep models to enhance the performance of detection and generalisation further. This problem is motivated by the need to achieve generalization against new and rapidly evolving attack variants. It is possible to extract identifiable features from presentation attack variants in order to detect them. However, new methods are needed to deal with emerging attacks and improve the generalization capability. This thesis examines the necessary measures to detect face presentation attacks in a more robust and generalised manner.
Citation
ABDULLAKUTTY, F.C. 2023. Unmasking the imposters: towards improving the generalisation of deep learning methods for face presentation attack detection. Robert Gordon University, PhD thesis. Hosted on OpenAIR [online]. Available from: https://doi.org/10.48526/rgu-wt-2270640
Thesis Type | Thesis |
---|---|
Deposit Date | Mar 12, 2024 |
Publicly Available Date | Mar 12, 2024 |
DOI | https://doi.org/10.48526/rgu-wt-2270640 |
Keywords | Facial recognition; Presentation attacks; Systems security; Deep learning; Neural networks |
Public URL | https://rgu-repository.worktribe.com/output/2270640 |
Award Date | Oct 31, 2023 |
Files
ABDULLAKUTTY 2023 Unmasking the imposters
(12.6 Mb)
PDF
Licence
https://creativecommons.org/licenses/by-nc/4.0/
Copyright Statement
© The Author.
You might also like
Fusion methods for face presentation attack detection.
(2022)
Journal Article
Deep transfer learning on the aggregated dataset for face presentation attack detection.
(2022)
Journal Article
Unmasking the imposters: task-specific feature learning for face presentation attack detection.
(2023)
Presentation / Conference Contribution
Decoding memes: a comprehensive analysis of late and early fusion models for explainable meme analysis.
(2024)
Presentation / Conference Contribution
Downloadable Citations
About OpenAIR@RGU
Administrator e-mail: publications@rgu.ac.uk
This application uses the following open-source libraries:
SheetJS Community Edition
Apache License Version 2.0 (http://www.apache.org/licenses/)
PDF.js
Apache License Version 2.0 (http://www.apache.org/licenses/)
Font Awesome
SIL OFL 1.1 (http://scripts.sil.org/OFL)
MIT License (http://opensource.org/licenses/mit-license.html)
CC BY 3.0 ( http://creativecommons.org/licenses/by/3.0/)
Powered by Worktribe © 2025
Advanced Search