Utama, Prasetya Ajie (2024)
Robustness of Pre-trained Language Models for Natural Language Understanding.
Technische Universität Darmstadt
doi: 10.26083/tuprints-00026582
Dissertation, Erstveröffentlichung, Verlagsversion
Kurzbeschreibung (Abstract)
Recent advances in neural network architectures and large-scale language model pretraining have enabled Natural Language Understanding (NLU) systems to surpass human-level performance on various benchmark datasets. However, a large body of work has revealed that NLU models are brittle against examples from outside of the training data distribution, which consequently limits their real-world application. This brittleness is mainly attributed to models exploiting spurious correlations in the training dataset. That is, models learn to use cues or shortcuts rather than robust features that are representative of the underlying task.
In this thesis, we present several methods to alleviate the effect of spurious correlation on the resulting NLU models. We attempt to improve the robustness against spurious correlation from several directions. Firstly, we address the issues in modeling methods that “debias” NLU models by reducing the incentives to learn non-robust features. We introduce a regularization method that uses the existing knowledge about spurious features’ characteristics to improve the out-of-distribution generalization without degrading the original performance on the standard evaluation. We further propose a strategy to maintain the effectiveness of the debiasing methods when the required prior knowledge is not available. Specifically, we introduce a self-debiasing framework that allows the identification of potentially biased examples that models should be disincentivized to exploit. Next, we also look at the inherent robustness that language models acquire during the pre-training on large text corpora. We show how task-specific fine-tuning can be destructive to such robustness and propose a novel regularizing approach to alleviate the degradation. Lastly, we tackle the issue of data augmentation approaches that aim to improve the robust performance of NLU models over downstream application tasks. We present a method to automatically generate diverse and naturalistic examples from which models can reliably learn the task.
In all task settings, we present in this thesis, models are evaluated against out-ofdistribution examples designed to penalize the reliance on spurious correlations. We measure the improvement in robustness by showing the increase in performance on these examples without the degradation of the existing standard evaluation. Overall, the work in this thesis demonstrates that we can still obtain robust NLU models using improved modeling and augmentation despite the presence of spurious correlations in the existing training resources.
Typ des Eintrags: | Dissertation | ||||
---|---|---|---|---|---|
Erschienen: | 2024 | ||||
Autor(en): | Utama, Prasetya Ajie | ||||
Art des Eintrags: | Erstveröffentlichung | ||||
Titel: | Robustness of Pre-trained Language Models for Natural Language Understanding | ||||
Sprache: | Englisch | ||||
Referenten: | Gurevych, Prof. Dr. Iryna ; Moosavi, Prof. Dr. Nafise Sadat ; Schwartz, Prof. Dr. Roy | ||||
Publikationsjahr: | 5 Februar 2024 | ||||
Ort: | Darmstadt | ||||
Kollation: | xi, 131 Seiten | ||||
Datum der mündlichen Prüfung: | 24 Oktober 2023 | ||||
DOI: | 10.26083/tuprints-00026582 | ||||
URL / URN: | https://tuprints.ulb.tu-darmstadt.de/26582 | ||||
Kurzbeschreibung (Abstract): | Recent advances in neural network architectures and large-scale language model pretraining have enabled Natural Language Understanding (NLU) systems to surpass human-level performance on various benchmark datasets. However, a large body of work has revealed that NLU models are brittle against examples from outside of the training data distribution, which consequently limits their real-world application. This brittleness is mainly attributed to models exploiting spurious correlations in the training dataset. That is, models learn to use cues or shortcuts rather than robust features that are representative of the underlying task. In this thesis, we present several methods to alleviate the effect of spurious correlation on the resulting NLU models. We attempt to improve the robustness against spurious correlation from several directions. Firstly, we address the issues in modeling methods that “debias” NLU models by reducing the incentives to learn non-robust features. We introduce a regularization method that uses the existing knowledge about spurious features’ characteristics to improve the out-of-distribution generalization without degrading the original performance on the standard evaluation. We further propose a strategy to maintain the effectiveness of the debiasing methods when the required prior knowledge is not available. Specifically, we introduce a self-debiasing framework that allows the identification of potentially biased examples that models should be disincentivized to exploit. Next, we also look at the inherent robustness that language models acquire during the pre-training on large text corpora. We show how task-specific fine-tuning can be destructive to such robustness and propose a novel regularizing approach to alleviate the degradation. Lastly, we tackle the issue of data augmentation approaches that aim to improve the robust performance of NLU models over downstream application tasks. We present a method to automatically generate diverse and naturalistic examples from which models can reliably learn the task. In all task settings, we present in this thesis, models are evaluated against out-ofdistribution examples designed to penalize the reliance on spurious correlations. We measure the improvement in robustness by showing the increase in performance on these examples without the degradation of the existing standard evaluation. Overall, the work in this thesis demonstrates that we can still obtain robust NLU models using improved modeling and augmentation despite the presence of spurious correlations in the existing training resources. |
||||
Alternatives oder übersetztes Abstract: |
|
||||
Status: | Verlagsversion | ||||
URN: | urn:nbn:de:tuda-tuprints-265828 | ||||
Sachgruppe der Dewey Dezimalklassifikatin (DDC): | 000 Allgemeines, Informatik, Informationswissenschaft > 004 Informatik 600 Technik, Medizin, angewandte Wissenschaften > 600 Technik 600 Technik, Medizin, angewandte Wissenschaften > 620 Ingenieurwissenschaften und Maschinenbau |
||||
Fachbereich(e)/-gebiet(e): | 20 Fachbereich Informatik 20 Fachbereich Informatik > Ubiquitäre Wissensverarbeitung |
||||
Hinterlegungsdatum: | 05 Feb 2024 13:03 | ||||
Letzte Änderung: | 05 Mär 2024 14:41 | ||||
PPN: | |||||
Referenten: | Gurevych, Prof. Dr. Iryna ; Moosavi, Prof. Dr. Nafise Sadat ; Schwartz, Prof. Dr. Roy | ||||
Datum der mündlichen Prüfung / Verteidigung / mdl. Prüfung: | 24 Oktober 2023 | ||||
Export: | |||||
Suche nach Titel in: | TUfind oder in Google |
Frage zum Eintrag |
Optionen (nur für Redakteure)
Redaktionelle Details anzeigen |