Buchmann, Jan ; Eichler, Max ; Bodensohn, Jan-Micha ; Kuznetsov, Ilia ; Gurevych, Iryna (2024)
Document Structure in Long Document Transformers.
18th Conference of the European Chapter of the Association for Computational Linguistics. St. Julian's, Malta (17.-22.03.2024)
Konferenzveröffentlichung, Bibliographie
Kurzbeschreibung (Abstract)
Long documents often exhibit structure with hierarchically organized elements of different functions, such as section headers and paragraphs. Despite the omnipresence of document structure, its role in natural language processing (NLP) remains opaque. Do long-document Transformer models acquire an internal representation of document structure during pre-training? How can structural information be communicated to a model after pre-training, and how does it influence downstream performance? To answer these questions, we develop a novel suite of probing tasks to assess structure-awareness of long-document Transformers, propose general-purpose structure infusion methods, and evaluate the effects of structure infusion on QASPER and Evidence Inference, two challenging long-document NLP tasks. Results on LED and LongT5 suggest that they acquire implicit understanding of document structure during pre-training, which can be further enhanced by structure infusion, leading to improved end-task performance. To foster research on the role of document structure in NLP modeling, we make our data and code publicly available.
Typ des Eintrags: | Konferenzveröffentlichung |
---|---|
Erschienen: | 2024 |
Autor(en): | Buchmann, Jan ; Eichler, Max ; Bodensohn, Jan-Micha ; Kuznetsov, Ilia ; Gurevych, Iryna |
Art des Eintrags: | Bibliographie |
Titel: | Document Structure in Long Document Transformers |
Sprache: | Englisch |
Publikationsjahr: | 23 März 2024 |
Ort: | St. Julian's, Malta |
Verlag: | ACL |
Buchtitel: | Proceedings of the 18th Conference of the European Chapter of the Association for Computational Linguistics (Volume 1: Long Papers) |
Veranstaltungstitel: | 18th Conference of the European Chapter of the Association for Computational Linguistics |
Veranstaltungsort: | St. Julian's, Malta |
Veranstaltungsdatum: | 17.-22.03.2024 |
URL / URN: | https://aclanthology.org/2024.eacl-long.64/ |
Kurzbeschreibung (Abstract): | Long documents often exhibit structure with hierarchically organized elements of different functions, such as section headers and paragraphs. Despite the omnipresence of document structure, its role in natural language processing (NLP) remains opaque. Do long-document Transformer models acquire an internal representation of document structure during pre-training? How can structural information be communicated to a model after pre-training, and how does it influence downstream performance? To answer these questions, we develop a novel suite of probing tasks to assess structure-awareness of long-document Transformers, propose general-purpose structure infusion methods, and evaluate the effects of structure infusion on QASPER and Evidence Inference, two challenging long-document NLP tasks. Results on LED and LongT5 suggest that they acquire implicit understanding of document structure during pre-training, which can be further enhanced by structure infusion, leading to improved end-task performance. To foster research on the role of document structure in NLP modeling, we make our data and code publicly available. |
Freie Schlagworte: | UKP_p_InterText |
Fachbereich(e)/-gebiet(e): | 20 Fachbereich Informatik 20 Fachbereich Informatik > Ubiquitäre Wissensverarbeitung |
Hinterlegungsdatum: | 12 Apr 2024 11:05 |
Letzte Änderung: | 08 Aug 2024 08:02 |
PPN: | 520432177 |
Export: | |
Suche nach Titel in: | TUfind oder in Google |
Frage zum Eintrag |
Optionen (nur für Redakteure)
Redaktionelle Details anzeigen |