TU Darmstadt / ULB / TUbiblio

Privacy-Preserving Graph Convolutional Networks for Text Classification

Igamberdiev, Timour ; Habernal, Ivan (2022)
Privacy-Preserving Graph Convolutional Networks for Text Classification.
13th Conference on Language Resources and Evaluation. Marseille, France (20.06.2022-25.06.2022)
Konferenzveröffentlichung, Bibliographie

Kurzbeschreibung (Abstract)

Graph convolutional networks (GCNs) are a powerful architecture for representation learning on documents that naturally occur as graphs, e.g., citation or social networks. However, sensitive personal information, such as documents with people’s profiles or relationships as edges, are prone to privacy leaks, as the trained model might reveal the original input. Although differential privacy (DP) offers a well-founded privacy-preserving framework, GCNs pose theoretical and practical challenges due to their training specifics. We address these challenges by adapting differentially-private gradient-based training to GCNs and conduct experiments using two optimizers on five NLP datasets in two languages. We propose a simple yet efficient method based on random graph splits that not only improves the baseline privacy bounds by a factor of 2.7 while retaining competitive F1 scores, but also provides strong privacy guarantees of ε = 1.0. We show that, under certain modeling choices, privacy-preserving GCNs perform up to 90% of their non-private variants, while formally guaranteeing strong privacy measures.

Typ des Eintrags: Konferenzveröffentlichung
Erschienen: 2022
Autor(en): Igamberdiev, Timour ; Habernal, Ivan
Art des Eintrags: Bibliographie
Titel: Privacy-Preserving Graph Convolutional Networks for Text Classification
Sprache: Englisch
Publikationsjahr: 29 Juni 2022
Verlag: European Language Resources Association (ELRA)
Buchtitel: LREC 2022 Conference Proceedings
Veranstaltungstitel: 13th Conference on Language Resources and Evaluation
Veranstaltungsort: Marseille, France
Veranstaltungsdatum: 20.06.2022-25.06.2022
URL / URN: http://www.lrec-conf.org/proceedings/lrec2022/index.html
Zugehörige Links:
Kurzbeschreibung (Abstract):

Graph convolutional networks (GCNs) are a powerful architecture for representation learning on documents that naturally occur as graphs, e.g., citation or social networks. However, sensitive personal information, such as documents with people’s profiles or relationships as edges, are prone to privacy leaks, as the trained model might reveal the original input. Although differential privacy (DP) offers a well-founded privacy-preserving framework, GCNs pose theoretical and practical challenges due to their training specifics. We address these challenges by adapting differentially-private gradient-based training to GCNs and conduct experiments using two optimizers on five NLP datasets in two languages. We propose a simple yet efficient method based on random graph splits that not only improves the baseline privacy bounds by a factor of 2.7 while retaining competitive F1 scores, but also provides strong privacy guarantees of ε = 1.0. We show that, under certain modeling choices, privacy-preserving GCNs perform up to 90% of their non-private variants, while formally guaranteeing strong privacy measures.

Fachbereich(e)/-gebiet(e): 20 Fachbereich Informatik
20 Fachbereich Informatik > Ubiquitäre Wissensverarbeitung
Hinterlegungsdatum: 13 Jul 2022 06:54
Letzte Änderung: 21 Nov 2022 14:54
PPN: 501817328
Export:
Suche nach Titel in: TUfind oder in Google
Frage zum Eintrag Frage zum Eintrag

Optionen (nur für Redakteure)
Redaktionelle Details anzeigen Redaktionelle Details anzeigen