From 2a17fa06f8cc0580f423aab72d657917febb46f7 Mon Sep 17 00:00:00 2001 From: AlvinYu <125203137+AlvinYu025@users.noreply.github.com> Date: Mon, 26 Jun 2023 17:35:23 +0800 Subject: [PATCH] Update README.md --- README.md | 12 +++++------- 1 file changed, 5 insertions(+), 7 deletions(-) diff --git a/README.md b/README.md index 283831f..8321c17 100644 --- a/README.md +++ b/README.md @@ -327,12 +327,11 @@ CVPR 2023 - Re-thinking Model Inversion Attacks Against Deep Neural Networks ## Graph learning domain | Year | Title | Adversarial Knowledge | Venue | Paper Link | Code Link | | ---- | ----- | -------------------- | ----- | ---------- | --------- | -| 2023 | NetGuard: Protecting Commercial Web APIs from Model Inversion Atacks using GAN-generated Fake Samples | white-box | ACM | [Paper](https://dl.acm.org/doi/pdf/10.1145/3543507.3583224) | | | 2020 | Improving Robustness to Model Inversion Attacks via Mutual Information Regularization | black & white-box | AAAI | [Paper](https://arxiv.org/pdf/2009.05241v1.pdf) | | +| 2020 | Reducing Risk of Model Inversion Using Privacy-Guided Training | black & white-box | Arxiv | [Paper](https://arxiv.org/pdf/2006.15877.pdf) | | | 2021 | A Survey on Gradient Inversion: Attacks, Defenses and Future Directions | white-box | IJCAI | [Paper](https://arxiv.org/pdf/2206.07284.pdf) | | | 2021 | GraphMI: Extracting Private Graph Data from Graph Neural Networks | white-box | IJCAI | [Paper](https://arxiv.org/pdf/2106.02820v1.pdf) | [code](https://github.com/zaixizhang/GraphMI) | -| 2021 | NetFense: Adversarial Defenses against Privacy Attacks on Neural Networks for Graph Data | black-box | IEEE | [Paper](https://arxiv.org/pdf/2106.11865.pdf) | [code](https://github.com/ICHproject/NetFense) | -| 2020 | Reducing Risk of Model Inversion Using Privacy-Guided Training | black & white-box | Arxiv | [Paper](https://arxiv.org/pdf/2006.15877.pdf) | | +| 2021 | NetFense: Adversarial Defenses against Privacy Attacks on Neural Networks for Graph Data | black-box | ICDE | [Paper](https://arxiv.org/pdf/2106.11865.pdf) | [code](https://github.com/ICHproject/NetFense) | | 2022 | A Comprehensive Survey on Trustworthy Graph Neural Networks: Privacy, Robustness, Fairness, and Explainability | black & white-box | Arxiv | [Paper](https://arxiv.org/pdf/2204.08570.pdf) | | USENIX Security 2020 - Stealing Links from Graph Neural Networks. @@ -400,11 +399,10 @@ ICML 2023 - On Strengthening and Defending Graph Reconstruction Attack with Mark | Year | Title | Adversarial Knowledge | Venue | Paper Link | Code Link | | ---- | ----- | -------------------- | ----- | ---------- | --------- | | 2020 | Extracting Training Data from Large Language Models | black-box | USENIX Security | [Paper](https://arxiv.org/pdf/2012.07805.pdf) | [code](https://arxiv.org/pdf/2012.07805.pdf) | -| 2020 | Privacy Risks of General-Purpose Language Models | black & white-box | IEEE | [Paper](https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=9152761) | | -| 2022 | Defending against Reconstruction Attacks with Rényi Differential Privacy | white-box | Arxiv | [Paper](https://arxiv.org/pdf/2202.07623.pdf) | | -| 2018 | Privacy-preserving Neural Representations of Text | white-box | Arxiv | [Paper](https://arxiv.org/pdf/1808.09408.pdf) | [code](https://github.com/mcoavoux/pnet) | +| 2020 | Privacy Risks of General-Purpose Language Models | black & white-box | S&P | [Paper](https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=9152761) | | +| 2018 | Privacy-preserving Neural Representations of Text | white-box | EMNLP | [Paper](https://arxiv.org/pdf/1808.09408.pdf) | [code](https://github.com/mcoavoux/pnet) | | 2021 | TAG: Gradient Attack on Transformer-based Language Models | white-box | EMNLP | [Paper](https://arxiv.org/pdf/2103.06819.pdf) | | -| 2020 | Information Leakage in Embedding Models | black & white-box | ACM | [Paper](https://arxiv.org/pdf/2004.00053.pdf) | | +| 2020 | Information Leakage in Embedding Models | black & white-box | CCS | [Paper](https://arxiv.org/pdf/2004.00053.pdf) | | | 2022 | Text Revealer: Private Text Reconstruction via Model Inversion Attacks against Transformers| white-box| Arxiv | [Paper](https://arxiv.org/pdf/2209.10505.pdf) | | CCS 2020 - Information Leakage in Embedding Models.