TU Darmstadt / ULB / TUbiblio

Document Structure in Long Document Transformers

Buchmann, Jan ; Eichler, Max ; Bodensohn, Jan-Micha ; Kuznetsov, Ilia ; Gurevych, Iryna (2024)
Document Structure in Long Document Transformers.
18th Conference of the European Chapter of the Association for Computational Linguistics. St. Julian's, Malta (17.-22.03.2024)
Konferenzveröffentlichung, Bibliographie

Kurzbeschreibung (Abstract)

Long documents often exhibit structure with hierarchically organized elements of different functions, such as section headers and paragraphs. Despite the omnipresence of document structure, its role in natural language processing (NLP) remains opaque. Do long-document Transformer models acquire an internal representation of document structure during pre-training? How can structural information be communicated to a model after pre-training, and how does it influence downstream performance? To answer these questions, we develop a novel suite of probing tasks to assess structure-awareness of long-document Transformers, propose general-purpose structure infusion methods, and evaluate the effects of structure infusion on QASPER and Evidence Inference, two challenging long-document NLP tasks. Results on LED and LongT5 suggest that they acquire implicit understanding of document structure during pre-training, which can be further enhanced by structure infusion, leading to improved end-task performance. To foster research on the role of document structure in NLP modeling, we make our data and code publicly available.

Typ des Eintrags: Konferenzveröffentlichung
Erschienen: 2024
Autor(en): Buchmann, Jan ; Eichler, Max ; Bodensohn, Jan-Micha ; Kuznetsov, Ilia ; Gurevych, Iryna
Art des Eintrags: Bibliographie
Titel: Document Structure in Long Document Transformers
Sprache: Englisch
Publikationsjahr: 23 März 2024
Ort: St. Julian's, Malta
Verlag: ACL
Buchtitel: Proceedings of the 18th Conference of the European Chapter of the Association for Computational Linguistics (Volume 1: Long Papers)
Veranstaltungstitel: 18th Conference of the European Chapter of the Association for Computational Linguistics
Veranstaltungsort: St. Julian's, Malta
Veranstaltungsdatum: 17.-22.03.2024
URL / URN: https://aclanthology.org/2024.eacl-long.64/
Kurzbeschreibung (Abstract):

Long documents often exhibit structure with hierarchically organized elements of different functions, such as section headers and paragraphs. Despite the omnipresence of document structure, its role in natural language processing (NLP) remains opaque. Do long-document Transformer models acquire an internal representation of document structure during pre-training? How can structural information be communicated to a model after pre-training, and how does it influence downstream performance? To answer these questions, we develop a novel suite of probing tasks to assess structure-awareness of long-document Transformers, propose general-purpose structure infusion methods, and evaluate the effects of structure infusion on QASPER and Evidence Inference, two challenging long-document NLP tasks. Results on LED and LongT5 suggest that they acquire implicit understanding of document structure during pre-training, which can be further enhanced by structure infusion, leading to improved end-task performance. To foster research on the role of document structure in NLP modeling, we make our data and code publicly available.

Freie Schlagworte: UKP_p_InterText
Fachbereich(e)/-gebiet(e): 20 Fachbereich Informatik
20 Fachbereich Informatik > Ubiquitäre Wissensverarbeitung
Hinterlegungsdatum: 12 Apr 2024 11:05
Letzte Änderung: 12 Apr 2024 11:05
PPN:
Export:
Suche nach Titel in: TUfind oder in Google
Frage zum Eintrag Frage zum Eintrag

Optionen (nur für Redakteure)
Redaktionelle Details anzeigen Redaktionelle Details anzeigen