Continuous Spiking Graph Neural Networks
CoRR(2024)
摘要
Continuous graph neural networks (CGNNs) have garnered significant attention
due to their ability to generalize existing discrete graph neural networks
(GNNs) by introducing continuous dynamics. They typically draw inspiration from
diffusion-based methods to introduce a novel propagation scheme, which is
analyzed using ordinary differential equations (ODE). However, the
implementation of CGNNs requires significant computational power, making them
challenging to deploy on battery-powered devices. Inspired by recent spiking
neural networks (SNNs), which emulate a biological inference process and
provide an energy-efficient neural architecture, we incorporate the SNNs with
CGNNs in a unified framework, named Continuous Spiking Graph Neural Networks
(COS-GNN). We employ SNNs for graph node representation at each time step,
which are further integrated into the ODE process along with time. To enhance
information preservation and mitigate information loss in SNNs, we introduce
the high-order structure of COS-GNN, which utilizes the second-order ODE for
spiking representation and continuous propagation. Moreover, we provide the
theoretical proof that COS-GNN effectively mitigates the issues of exploding
and vanishing gradients, enabling us to capture long-range dependencies between
nodes. Experimental results on graph-based learning tasks demonstrate the
effectiveness of the proposed COS-GNN over competitive baselines.
更多查看译文
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要