Subgraph Pooling: Tackling Negative Transfer on Graphs
arxiv(2024)
摘要
Transfer learning aims to enhance performance on a target task by using
knowledge from related tasks. However, when the source and target tasks are not
closely aligned, it can lead to reduced performance, known as negative
transfer. Unlike in image or text data, we find that negative transfer could
commonly occur in graph-structured data, even when source and target graphs
have semantic similarities. Specifically, we identify that structural
differences significantly amplify the dissimilarities in the node embeddings
across graphs. To mitigate this, we bring a new insight in this paper: for
semantically similar graphs, although structural differences lead to
significant distribution shift in node embeddings, their impact on subgraph
embeddings could be marginal. Building on this insight, we introduce Subgraph
Pooling (SP) by aggregating nodes sampled from a k-hop neighborhood and
Subgraph Pooling++ (SP++) by a random walk, to mitigate the impact of graph
structural differences on knowledge transfer. We theoretically analyze the role
of SP in reducing graph discrepancy and conduct extensive experiments to
evaluate its superiority under various settings. The proposed SP methods are
effective yet elegant, which can be easily applied on top of any backbone Graph
Neural Networks (GNNs). Our code and data are available at:
https://github.com/Zehong-Wang/Subgraph-Pooling.
更多查看译文
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要