| 研究生: |
葉嘉浤 Yeh, Chia-Hung |
|---|---|
| 論文名稱: |
學習穩健圖神經網路以應對真偽難辨且多元的圖對抗攻擊 Learning Robust Graph Neural Networks to Combat Diverse and Elusive Graph Adversarial Attacks |
| 指導教授: |
李政德
Li, Cheng-Te |
| 共同指導教授: |
張欣民
Chang, Hsing-Ming |
| 學位類別: |
碩士 Master |
| 系所名稱: |
管理學院 - 數據科學研究所 Institute of Data Science |
| 論文出版年: | 2024 |
| 畢業學年度: | 112 |
| 語文別: | 英文 |
| 論文頁數: | 44 |
| 中文關鍵詞: | 深度學習 、圖神經網路 、穩健性 |
| 外文關鍵詞: | Deep Learning, Graph Neural Networks, Robustness |
| 相關次數: | 點閱:55 下載:0 |
| 分享至: |
| 查詢本校圖書館目錄 查詢臺灣博碩士論文知識加值系統 勘誤回報 |
圖神經網絡展示了它們在建模圖結構方面的強大能力,並在各種領域任務中達到了卓越的性能。然而,在真實世界中的圖往往存在雜訊,並且帶有標籤的節點數量非常少,這會大幅的降低圖神經網路模型在下游任務的表現,使得圖神經網路難以運用在真實情境中。過去的許多研究假定輸入的節點特徵是乾淨的,並以此去學習對有雜訊的圖結構做重建,但輸入的節點特徵也帶有雜訊時,就不適合使用這些方法。因此,本研究提出一個穩健性的圖神經網路模型,可以對有雜訊的圖結構與節點特徵做修正,並在有標籤的節點數量非常少的情形下,仍然能為每個節點做出正確的預測結果。首先,我們使用拉普拉斯平滑化,使得原本給定的具有雜訊的節點特徵在經過平滑化以後,可以較接近沒有雜訊的節點特徵經過平滑化後的特徵分布,並使相鄰節點的距離會更接近。然後,我們會利用平滑過的節點特徵,輸入到一個同質性增強編碼器,進一步使得原本具有高相似性的的節點會學到更相似的特徵表示,反之則會學到越不相似的特徵表示。接著,我們會將學到的節點特徵表示輸入到一個用於執行鏈接預測的編碼器,依據每個節點特徵表示計算彼此的相似度作為節點對之間的邊權重,以此來降低或消除被視為雜訊的邊並形成更多鏈接,使訊息能更有效的傳遞,最後,我們將學習到的節點特徵表示與圖結構輸入到圖神經網路分類器,執行節點分類任務,並加入一項正則化,使得原本沒有標籤的節點可以直接被計算到目標函數中。在實驗階段,我們使用四個常見的學術引用網路資料集進行節點分類任務、模型穩健性分析、消融分析、超參數分析。實驗結果證明我們的方法在輸入帶有雜訊的圖結構與節點特徵時,相較於過去的方法,可以得到較高的準確率,表示我們提出的模型對於圖結構以及節點特徵的雜訊具有穩健性,並在標籤稀疏性的狀況下也能為每個節點做出正確的預測。
Graph Neural Networks (GNNs) have exhibited their capability on modeling graph, and reached epic performance on various domains. However, graphs in the real world often contain structure/feature noise and have few labeled nodes, which significantly diminishes the performance of GNNs in downstream tasks, making it difficult to apply GNNs in real-world scenarios. Previous studies assume that the input node features are clean and use this assumption to learn how to reconstruct graph. Nevertheless, when the node features are also noisy, these methods are no longer suitable. Therefore, we propose a Adversarial Robust GNN (ARGNN) that can correct noisy graph and node features, and make accurate predictions for each node with limited labeled nodes. First, we use Laplacian smoothing filter to denoise the given noisy node features and makes the smoothed node features more closer to the smoothed clean node features, while enhance the homophily for neighboring nodes. Then, we utilize the smoothed node feature as input to a homophily-augmented encoder, which learns more similar node embeddings for nodes with originally similar features and more dissimilar node embeddings for nodes with dissimilar features. After obtaining the learned node embedding matrix, we feed it into a link predictor to learn different edge weights based on the pairwise feature similarity of each node pair to reduce/remove the noisy edges and create more connections for more effectively message-passing. Finally, we input the learned node embeddings and the graph structure to a GNN classifier for node classification task, and introduce a regularization term which makes the unlabeled nodes can be directly computed into the objective function. In experiment phase, we conduct node classification task, model robustness analysis, ablation study, and hyperparameter analysis on four citation network dataset. The experiment results demonstrate that our method achieves higher accuracy compared to state-of-the-art approaches when the input graph and node features contain noise, meaning that our proposed framework is robustness to the structural noise and feature noise, and can make correct prediction for each node with limited labeled nodes.
[1] Amr Ahmed, Nino Shervashidze, Shravan Narayanamurthy, Vanja Josifovski, and Alexander J Smola. Distributed large-scale natural graph factorization. In Proceedings of the 22nd international conference on World Wide Web, pages 37–48, 2013.
[2] AK Awasthi, Arun Kumar Garov, Minakshi Sharma, and Mrigank Sinha. Gnn model based on node classification forecasting in social network. In 2023 International Conference on Artificial Intelligence and Smart Communication (AISC), pages 1039–1043. IEEE, 2023.
[3] Lei Bai, Lina Yao, Can Li, Xianzhi Wang, and Can Wang. Adaptive graph convolutional recurrent network for traffic forecasting. Advances in neural information processing systems, 33:17804–17815, 2020.
[4] Mikhail Belkin and Partha Niyogi. Laplacian eigenmaps for dimensionality reduction and data representation. Neural computation, 15(6):1373–1396, 2003.
[5] Yoshua Bengio, Jérôme Louradour, Ronan Collobert, and Jason Weston. Curriculum learning. In Proceedings of the 26th annual international conference on machine learning, pages 41–48, 2009.
[6] Shaosheng Cao, Wei Lu, and Qiongkai Xu. Grarep: Learning graph representations with global structural information. In Proceedings of the 24th ACM international on conference on information and knowledge management, pages 891–900, 2015.
[7] Jianlong Chang, Lingfeng Wang, Gaofeng Meng, Shiming Xiang, and Chunhong Pan. Deep adaptive image clustering. In Proceedings of the IEEE international conference on computer vision, pages 5879–5887, 2017.
[8] Ganqu Cui, Jie Zhou, Cheng Yang, and Zhiyuan Liu. Adaptive graph encoder for attributed graph embedding. In Proceedings of the 26th ACM SIGKDD international conference on knowledge discovery & data mining, pages 976–985, 2020.
[9] Enyan Dai, Charu Aggarwal, and Suhang Wang. Nrgnn: Learning a label noise resistant graph neural network on sparsely and noisily labeled graphs. In Proceedings of the 27th ACM SIGKDD conference on knowledge discovery & data mining, pages 227–236, 2021.
[10] Enyan Dai, Wei Jin, Hui Liu, and Suhang Wang. Towards robust graph neural networks for noisy graphs with sparse labels. In Proceedings of the Fifteenth ACM International Conference on Web Search and Data Mining, pages 181–191, 2022.
[11] Hanjun Dai, Hui Li, Tian Tian, Xin Huang, Lin Wang, Jun Zhu, and Le Song. Adversarial attack on graph structured data. In International conference on machine learning, pages 1115–1124. PMLR, 2018.
[12] Negin Entezari, Saba A Al-Sayouri, Amirali Darvishzadeh, and Evangelos E Papalexakis. All you need is low (rank) defending against adversarial attacks on graphs. In Proceedings of the 13th international conference on web search and data mining, pages 169–177, 2020.
[13] Francesco Folino, Gianluigi Folino, Massimo Guarascio, Luigi Pontieri, and Paolo Zicari. Towards data-and compute-efficient fake-news detection: An approach combining active learning and pre-trained language models. SN ComputerScience, 5(5):470, 2024.
[14] Aditya Grover and Jure Leskovec. node2vec: Scalable feature learning for networks. In Proceedings of the 22nd ACM SIGKDD international conference on Knowledge discovery and data mining, pages 855–864, 2016.
[15] Will Hamilton, Zhitao Ying, and Jure Leskovec. Inductive representation learning on large graphs. Advances in neural information processing systems, 30, 2017.
[16] Tai Hasegawa,Sukwon Yun,Xin Liu,YinJun Phua,and Tsuyoshi Murata. Degnn: Dual experts graph neural network handling both edge and node feature noise. In Pacific Asia Conference on Knowledge Discovery and Data Mining, pages 376–389. Springer, 2024.
[17] Roger A Horn and Charles R Johnson. Matrix analysis. Cambridge university press, 2012.
[18] Zhenyu Hou, Xiao Liu, Yukuo Cen, Yuxiao Dong, Hongxia Yang, Chunjie Wang, and Jie Tang. Graphmae: Self-supervised masked graph autoencoders. In Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, pages 594–604, 2022.
[19] Wei Jin, Tyler Derr, Yiqi Wang, Yao Ma, Zitao Liu, and Jiliang Tang. Node similarity preserving graph convolutional networks. In Proceedings of the 14th ACM international conference on web search and data mining, pages 148–156, 2021.
[20] Wei Jin, Yao Ma, Xiaorui Liu, Xianfeng Tang, Suhang Wang, and Jiliang Tang. Graph structure learning for robust graph neural networks. In Proceedings of the 26th ACM SIGKDD international conference on knowledge discovery & data mining, pages 66–74, 2020.
[21] Thomas N Kipf and Max Welling. Semi-supervised classification with graph convolutional networks. arXiv preprint arXiv:1609.02907, 2016.
[22] Thomas N Kipf and Max Welling. Variational graph auto-encoders. arXiv preprint arXiv:1611.07308, 2016.
[23] Chunyan Li, Wei Wei, Jin Li, Junfeng Yao, Xiangxiang Zeng, and Zhihan Lv. 3dmol-net: learn 3d molecular representation using adaptive graph convolutional network based on rotation invariance. IEEE journal of biomedical and health informatics, 26(10):5044–5054, 2021.
[24] Qimai Li, Zhichao Han, and Xiao-Ming Wu. Deeper insights into graph convolutional networks for semi-supervised learning, 2018.
[25] Andrew Kachites McCallum, Kamal Nigam, Jason Rennie, and Kristie Seymore. Automating the construction of internet portals with machine learning. Information Retrieval, 3:127–163, 2000.
[26] Mingdong Ou, Peng Cui, Jian Pei, Ziwei Zhang, and Wenwu Zhu. Asymmetric transitivity preserving graph embedding. In Proceedings of the 22nd ACM SIGKDD international conference on Knowledge discovery and data mining, pages 1105–1114, 2016.
[27] Bryan Perozzi, Rami Al-Rfou, and Steven Skiena. Deepwalk: Online learning of social representations. In Proceedings of the 20th ACM SIGKDD international conference on Knowledge discovery and data mining, pages 701–710, 2014.
[28] Prithviraj Sen, Galileo Namata, Mustafa Bilgic, Lise Getoor, Brian Galligher, and Tina Eliassi-Rad. Collective classification in network data. AI magazine, 29(3):93–93, 2008.
[29] Yuyol Shin and Yoonjin Yoon. Incorporating dynamicity of transportation network with multi-weight traffic graph convolutional network for traffic forecasting. IEEE Transactions on Intelligent Transportation Systems, 23(3):2082–2092, 2020.
[30] Gabriel Taubin. A signal processing approach to fair surface design. In Proceedings of the 22nd annual conference on Computer graphics and interactive techniques, pages 351–358, 1995.
[31] Petar Veličković, Guillem Cucurull, Arantxa Casanova, Adriana Romero, Pietro Lio, and Yoshua Bengio. Graphattention networks. arXiv preprint arXiv:1710.10903, 2017.
[32] Hongwei Wang, Jia Wang, Jialin Wang, Miao Zhao, Weinan Zhang, Fuzheng Zhang, Xing Xie, and Minyi Guo. Graphgan: Graph representation learning with generative adversarial nets. In Proceedings of the AAAI conference on artificial intelligence, volume 32, 2018.
[33] Xiao Wang, Houye Ji, Chuan Shi, Bai Wang, Yanfang Ye, Peng Cui, and Philip S Yu. Heterogeneous graph attention network. In The world wide web conference, pages 2022–2032, 2019.
[34] Huijun Wu, Chen Wang, Yuriy Tyshetskiy, Andrew Docherty, Kai Lu, and Liming Zhu. Adversarial examples on graph data: Deep insights into attack and defense. arXiv preprint arXiv:1903.01610, 2019.
[35] Junliang Yu, Hongzhi Yin, Xin Xia, Tong Chen, Lizhen Cui, and Quoc Viet Hung Nguyen. Are graph augmentations necessary? simple graph contrastive learning for recommendation. In Proceedings of the 45th international ACM SIGIR conference on research and development in information retrieval, pages 1294–1303, 2022.
[36] Xiang Zhang and Marinka Zitnik. Gnnguard: Defending graph neural networks against adversarial attacks. Advances in neural information processing systems, 33:9263–9275, 2020.
[37] Dingyuan Zhu, Ziwei Zhang, Peng Cui, and Wenwu Zhu. Robust graph convolutional networks against adversarial attacks. In Proceedings of the 25th ACM SIGKDD international conference on knowledge discovery & data mining, pages 1399–1407, 2019.
[38] Xiaojin ZhuЃ and Zoubin GhahramaniЃн. Learning from labeled and unlabeled data with label propagation. ProQuest number: information to all users, 2002.
[39] Daniel Zügner, Amir Akbarnejad, and Stephan Günnemann. Adversarial attacks on neural networks for graph data. In Proceedings of the 24th ACM SIGKDD international conference on knowledge discovery & data mining, pages 2847–2856, 2018.
[40] Daniel Zügner and Stephan Günnemann. Adversarial attacks on graph neural networks via meta learning. ArXiv, abs/1902.08412, 2019.
校內:2029-08-21公開