Igamberdiev, Timour ; Habernal, Ivan (2022)
Privacy-Preserving Graph Convolutional Networks for Text Classification.
13th Conference on Language Resources and Evaluation. Marseille, France (20.06.2022-25.06.2022)
Konferenzveröffentlichung, Bibliographie
Kurzbeschreibung (Abstract)
Graph convolutional networks (GCNs) are a powerful architecture for representation learning on documents that naturally occur as graphs, e.g., citation or social networks. However, sensitive personal information, such as documents with people’s profiles or relationships as edges, are prone to privacy leaks, as the trained model might reveal the original input. Although differential privacy (DP) offers a well-founded privacy-preserving framework, GCNs pose theoretical and practical challenges due to their training specifics. We address these challenges by adapting differentially-private gradient-based training to GCNs and conduct experiments using two optimizers on five NLP datasets in two languages. We propose a simple yet efficient method based on random graph splits that not only improves the baseline privacy bounds by a factor of 2.7 while retaining competitive F1 scores, but also provides strong privacy guarantees of ε = 1.0. We show that, under certain modeling choices, privacy-preserving GCNs perform up to 90% of their non-private variants, while formally guaranteeing strong privacy measures.
Typ des Eintrags: | Konferenzveröffentlichung |
---|---|
Erschienen: | 2022 |
Autor(en): | Igamberdiev, Timour ; Habernal, Ivan |
Art des Eintrags: | Bibliographie |
Titel: | Privacy-Preserving Graph Convolutional Networks for Text Classification |
Sprache: | Englisch |
Publikationsjahr: | 29 Juni 2022 |
Verlag: | European Language Resources Association (ELRA) |
Buchtitel: | LREC 2022 Conference Proceedings |
Veranstaltungstitel: | 13th Conference on Language Resources and Evaluation |
Veranstaltungsort: | Marseille, France |
Veranstaltungsdatum: | 20.06.2022-25.06.2022 |
URL / URN: | http://www.lrec-conf.org/proceedings/lrec2022/index.html |
Zugehörige Links: | |
Kurzbeschreibung (Abstract): | Graph convolutional networks (GCNs) are a powerful architecture for representation learning on documents that naturally occur as graphs, e.g., citation or social networks. However, sensitive personal information, such as documents with people’s profiles or relationships as edges, are prone to privacy leaks, as the trained model might reveal the original input. Although differential privacy (DP) offers a well-founded privacy-preserving framework, GCNs pose theoretical and practical challenges due to their training specifics. We address these challenges by adapting differentially-private gradient-based training to GCNs and conduct experiments using two optimizers on five NLP datasets in two languages. We propose a simple yet efficient method based on random graph splits that not only improves the baseline privacy bounds by a factor of 2.7 while retaining competitive F1 scores, but also provides strong privacy guarantees of ε = 1.0. We show that, under certain modeling choices, privacy-preserving GCNs perform up to 90% of their non-private variants, while formally guaranteeing strong privacy measures. |
Fachbereich(e)/-gebiet(e): | 20 Fachbereich Informatik 20 Fachbereich Informatik > Ubiquitäre Wissensverarbeitung |
Hinterlegungsdatum: | 13 Jul 2022 06:54 |
Letzte Änderung: | 21 Nov 2022 14:54 |
PPN: | 501817328 |
Export: | |
Suche nach Titel in: | TUfind oder in Google |
Frage zum Eintrag |
Optionen (nur für Redakteure)
Redaktionelle Details anzeigen |