TU Darmstadt / ULB / TUbiblio

MixQuantBio: Towards extreme face and periocular recognition model compression with mixed-precision quantization

Kolf, Jan Niklas ; Elliesen, Jurek ; Damer, Naser ; Boutros, Fadi (2024)
MixQuantBio: Towards extreme face and periocular recognition model compression with mixed-precision quantization.
In: Engineering Applications of Artificial Intelligence
doi: 10.1016/j.engappai.2024.109114
Artikel, Bibliographie

Kurzbeschreibung (Abstract)

Current periocular and face recognition approaches utilize computationally costly deep neural networks, achieving notable recognition accuracies. Deploying such solutions in applications with limited computational resources requires minimizing their computational demand while maintaining similar recognition accuracies. Model compression techniques like model quantization can be used to reduce the computational costs of deep models. This approach is widely studied and applied to different machine-learning tasks, however it is understudied and investigated for biometrics. We propose in this work to reduce the computational cost of face and periocular recognition models using fixed- and mixed-precision model quantization. Specifically, we first quantize the full-precision models to fixed 8 and 6 bits, reducing the required memory footprint by 5x while maintaining, to a very large degree, the recognition accuracies. However, our achieved results demonstrated that by quantizing the models to extremely low b bits, e.g., below 6 bits, the accuracies significantly dropped, which motivated our investigation on mixed-precision quantization. Hence, we propose to utilize an iterative mixed-precision quantization scheme. In each iteration, the least important parameters are selected based on their weight magnitude and quantized to low b-bit precision and the model is fine-tuned. This approach is repeated until all parameters are quantized to low b-bit precision, achieving extreme reduction in memory footprint, e.g., 16x times, without significant loss in the model accuracies. The effectiveness of mixed- and fixed-precision quantization for biometric recognition models is studied and proved for two modalities, face and periocular, using three different deep network architectures and using different b bit precision.

Typ des Eintrags: Artikel
Erschienen: 2024
Autor(en): Kolf, Jan Niklas ; Elliesen, Jurek ; Damer, Naser ; Boutros, Fadi
Art des Eintrags: Bibliographie
Titel: MixQuantBio: Towards extreme face and periocular recognition model compression with mixed-precision quantization
Sprache: Englisch
Publikationsjahr: 2024
Titel der Zeitschrift, Zeitung oder Schriftenreihe: Engineering Applications of Artificial Intelligence
Band einer Reihe: 137
DOI: 10.1016/j.engappai.2024.109114
URL / URN: https://doi.org/10.1016/j.engappai.2024.109114
Kurzbeschreibung (Abstract):

Current periocular and face recognition approaches utilize computationally costly deep neural networks, achieving notable recognition accuracies. Deploying such solutions in applications with limited computational resources requires minimizing their computational demand while maintaining similar recognition accuracies. Model compression techniques like model quantization can be used to reduce the computational costs of deep models. This approach is widely studied and applied to different machine-learning tasks, however it is understudied and investigated for biometrics. We propose in this work to reduce the computational cost of face and periocular recognition models using fixed- and mixed-precision model quantization. Specifically, we first quantize the full-precision models to fixed 8 and 6 bits, reducing the required memory footprint by 5x while maintaining, to a very large degree, the recognition accuracies. However, our achieved results demonstrated that by quantizing the models to extremely low b bits, e.g., below 6 bits, the accuracies significantly dropped, which motivated our investigation on mixed-precision quantization. Hence, we propose to utilize an iterative mixed-precision quantization scheme. In each iteration, the least important parameters are selected based on their weight magnitude and quantized to low b-bit precision and the model is fine-tuned. This approach is repeated until all parameters are quantized to low b-bit precision, achieving extreme reduction in memory footprint, e.g., 16x times, without significant loss in the model accuracies. The effectiveness of mixed- and fixed-precision quantization for biometric recognition models is studied and proved for two modalities, face and periocular, using three different deep network architectures and using different b bit precision.

Freie Schlagworte: Biometrics, Face recognition, Machine learning, Deep learning
Fachbereich(e)/-gebiet(e): 20 Fachbereich Informatik
20 Fachbereich Informatik > Graphisch-Interaktive Systeme
Hinterlegungsdatum: 10 Sep 2024 13:52
Letzte Änderung: 23 Okt 2024 06:51
PPN: 522396151
Export:
Suche nach Titel in: TUfind oder in Google
Frage zum Eintrag Frage zum Eintrag

Optionen (nur für Redakteure)
Redaktionelle Details anzeigen Redaktionelle Details anzeigen