TU Darmstadt / ULB / TUbiblio

The Devil is in the Details: On Models and Training Regimes for Few-Shot Intent Classification

Mesgar, Mohsen ; Tran, Thy Thy ; Glavaš, Goran ; Gurevych, Iryna (2023)
The Devil is in the Details: On Models and Training Regimes for Few-Shot Intent Classification.
17th Conference of the European Chapter of the Association for Computational Linguistics. Dubrovnik, Croatia (02.05.2023-06.05.2023)
Conference or Workshop Item, Bibliographie

Abstract

In task-oriented dialog (ToD) new intents emerge on regular basis, with a handful of available utterances at best. This renders effective Few-Shot Intent Classification (FSIC) a central challenge for modular ToD systems. Recent FSIC methods appear to be similar: they use pretrained language models (PLMs) to encode utterances and predominantly resort to nearest-neighbor-based inference. However, they also differ in major components: they start from different PLMs, use different encoding architectures and utterance similarity functions, and adopt different training regimes.Coupling of these vital components together with the lack of informative ablations prevents the identification of factors that drive the (reported) FSIC performance. We propose a unified framework to evaluate these components along the following key dimensions:(1) Encoding architectures: Cross-Encoder vs Bi-Encoders;(2) Similarity function: Parameterized (i.e., trainable) vs non-parameterized; (3) Training regimes: Episodic meta-learning vs conventional (i.e., non-episodic) training. Our experimental results on seven FSIC benchmarks reveal three new important findings. First, the unexplored combination of cross-encoder architecture and episodic meta-learning consistently yields the best FSIC performance. Second, episodic training substantially outperforms its non-episodic counterpart. Finally, we show that splitting episodes into support and query sets has a limited and inconsistent effect on performance. Our findings show the importance of ablations and fair comparisons in FSIC. We publicly release our code and data.

Item Type: Conference or Workshop Item
Erschienen: 2023
Creators: Mesgar, Mohsen ; Tran, Thy Thy ; Glavaš, Goran ; Gurevych, Iryna
Type of entry: Bibliographie
Title: The Devil is in the Details: On Models and Training Regimes for Few-Shot Intent Classification
Language: English
Date: 2 May 2023
Publisher: ACL
Book Title: The 17th Conference of the European Chapter of the Association for Computational Linguistics - proceedings of the conference
Event Title: 17th Conference of the European Chapter of the Association for Computational Linguistics
Event Location: Dubrovnik, Croatia
Event Dates: 02.05.2023-06.05.2023
URL / URN: https://aclanthology.org/2023.eacl-main.135/
Abstract:

In task-oriented dialog (ToD) new intents emerge on regular basis, with a handful of available utterances at best. This renders effective Few-Shot Intent Classification (FSIC) a central challenge for modular ToD systems. Recent FSIC methods appear to be similar: they use pretrained language models (PLMs) to encode utterances and predominantly resort to nearest-neighbor-based inference. However, they also differ in major components: they start from different PLMs, use different encoding architectures and utterance similarity functions, and adopt different training regimes.Coupling of these vital components together with the lack of informative ablations prevents the identification of factors that drive the (reported) FSIC performance. We propose a unified framework to evaluate these components along the following key dimensions:(1) Encoding architectures: Cross-Encoder vs Bi-Encoders;(2) Similarity function: Parameterized (i.e., trainable) vs non-parameterized; (3) Training regimes: Episodic meta-learning vs conventional (i.e., non-episodic) training. Our experimental results on seven FSIC benchmarks reveal three new important findings. First, the unexplored combination of cross-encoder architecture and episodic meta-learning consistently yields the best FSIC performance. Second, episodic training substantially outperforms its non-episodic counterpart. Finally, we show that splitting episodes into support and query sets has a limited and inconsistent effect on performance. Our findings show the importance of ablations and fair comparisons in FSIC. We publicly release our code and data.

Uncontrolled Keywords: UKP_p_square,UKP_p_SERMAS
Divisions: 20 Department of Computer Science
20 Department of Computer Science > Ubiquitous Knowledge Processing
Date Deposited: 12 Jun 2023 12:29
Last Modified: 09 Aug 2023 12:42
PPN: 51046971X
Export:
Suche nach Titel in: TUfind oder in Google
Send an inquiry Send an inquiry

Options (only for editors)
Show editorial Details Show editorial Details